Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Two is better than one: Apologies from two robots are preferred

  • Yuka Okada ,

    Contributed equally to this work with: Yuka Okada, Mitsuhiko Kimoto, Takamasa Iio, Katsunori Shimohara, Masahiro Shiomi

    Roles Conceptualization, Data curation, Investigation, Methodology, Software, Writing – review & editing

    Affiliations Department of Agent Interaction Design Laboratory, Advanced Telecommunications Research Institute International, Kyoto, Japan, Department of Information Systems Design, Doshisha University, Kyoto, Japan

  • Mitsuhiko Kimoto ,

    Contributed equally to this work with: Yuka Okada, Mitsuhiko Kimoto, Takamasa Iio, Katsunori Shimohara, Masahiro Shiomi

    Roles Conceptualization, Formal analysis, Funding acquisition, Investigation, Methodology, Supervision, Validation, Writing – review & editing

    Affiliation Department of Agent Interaction Design Laboratory, Advanced Telecommunications Research Institute International, Kyoto, Japan

  • Takamasa Iio ,

    Contributed equally to this work with: Yuka Okada, Mitsuhiko Kimoto, Takamasa Iio, Katsunori Shimohara, Masahiro Shiomi

    Roles Conceptualization, Formal analysis, Funding acquisition, Methodology, Supervision, Validation, Writing – review & editing

    Affiliations Department of Agent Interaction Design Laboratory, Advanced Telecommunications Research Institute International, Kyoto, Japan, Faculty of Culture and Information Science, Doshisha University, Kyoto, Japan

  • Katsunori Shimohara ,

    Contributed equally to this work with: Yuka Okada, Mitsuhiko Kimoto, Takamasa Iio, Katsunori Shimohara, Masahiro Shiomi

    Roles Conceptualization, Project administration, Resources, Supervision, Writing – review & editing

    Affiliations Department of Agent Interaction Design Laboratory, Advanced Telecommunications Research Institute International, Kyoto, Japan, Department of Information Systems Design, Doshisha University, Kyoto, Japan

  • Masahiro Shiomi

    Contributed equally to this work with: Yuka Okada, Mitsuhiko Kimoto, Takamasa Iio, Katsunori Shimohara, Masahiro Shiomi

    Roles Conceptualization, Formal analysis, Funding acquisition, Methodology, Project administration, Resources, Supervision, Validation, Visualization, Writing – original draft, Writing – review & editing

    m-shiomi@atr.jp

    Affiliation Department of Agent Interaction Design Laboratory, Advanced Telecommunications Research Institute International, Kyoto, Japan

Abstract

Although the capabilities of service robots are increasing, avoiding any mistakes is difficult. Therefore, strategies for mitigating mistakes, such as apology behavior designs, are essential for service robots. Past studies reported that costly apology is perceived as more sincere than non-costly ones and more acceptable. To increase the apology cost in robot service situations, we thought that using multiple robots would increase the perceived costs in the of financial, physical, and time costs. Therefore, we focused on the number of robots who apologize for their mistakes as well as their individual, specific roles and behaviors during such apologies. We investigated the differences in perceived impressions toward apologies from two robots (the main robot that makes a mistake and apologizes and a sub-robot that also apologizes) and an apology from just one robot (only the main robot) through a web survey with 168 valid participants. The experiment results showed that the participants significantly preferred and positively evaluated apologies from two robots more than one robot in the context of forgiveness, negative word-of-mouth, trust, and intention to use. We also conducted another web survey with 430 valid participants to investigate the effects of different roles for the sub-robot: apologize-only, cleaning-up-only, and both actions. The experimental results showed that the participants significantly preferred and positively evaluated both actions in the context of forgiveness and reliable/competent perspectives.

Introduction

Service robots are becoming increasingly common in daily environments. During the COVID-19 pandemic, they conducted various daily activities to mitigate the risk of infection [1, 2]. In the human-robot interaction research field, robotics researchers have investigated the effectiveness of social robots in shopping malls [35], public transportation [68], science museums [911], and nursing homes [1214]. These studies described what services and systems were useful to support daily living before the pandemic. These long-term academic contributions have fueled recent advances in service robots in commercial contexts.

For such service robots for daily environments, robotics researchers should consider a situation where robots are not perfect. Since we human beings obviously make mistakes, we have several strategies to mitigate them, such as offering an apology as a “social lubricant” [15]. In fact, people prefer to interact with a company that apologizes compared to a company that does not [16]. Several studies investigated the effects of apologies in economic and business contexts [17, 18] because the cost perspective is one key factor for accepting them. A past study has already proved that apology-forgiveness mechanisms emerge if the apology is costly enough up to a certain threshold [17]. Moreover, other studies provided a computational model that can explain the effects between apology’s cost and the degree of commitment, and the relationship between them and the iterated Prosoner’s Dilmma [19, 20]. Another study reported the importance of such costly apologies, which tend to be perceived as more sincere than non-costly ones [18]. A recent study in human-computer interaction reported that a sincere apology increases feelings of forgiveness, perceived intelligence, and likeability toward agents [21]. These studies suggest that the concept of cost in apologies and their sincerity are essential for mitigating strategies for social robots.

Although robotics researchers did not clearly include a cost-like concept, they have already implemented behaviors that diminish the mistakes of service robots and lead to forgiveness [2229]. However, we note that the existing studies of robot mitigation strategies mainly focused on a situation where a single robot apologizes. In the field of human-robot interaction, several studies have investigated situations where multiple robots collaborate to provide services. Several studies reported the effectiveness of using multiple robots during interactions with people, such as rewarding [30, 31] and information-providing services [3234]. These studies reported that even though the amount of information is identical between multiple robots and a single robot, people prefer the former.

Based on these considerations, we are interested in whether using multiple robots when apologizing is as effective as other service contexts. We thought that using multiple robots for an apology might increase the perceived costs more than an apology from just one robot in the context of financial, physical, and time costs, creating a situation that leads people to absolve a robot for its failures and has positive impressions about it.

We believe that apologies from multiple robots are a more prudent approach than a single apology from just one robot: e.g., the main robot that apologizes for its mistake (Fig 1, left) and a sub-robot that also apologizes (Fig 1, right)). Investigating the effectiveness of multiple-robot-apologizing behaviors will provide useful knowledge to design such mitigating strategies for service robots in actual environments. Based on these considerations, we address the following research question:

  1. Research question 1: Do people prefer apologies from two robots or just one?
thumbnail
Fig 1. A scene of apology.

Apology from one robot (left) and apologies from two (right).

https://doi.org/10.1371/journal.pone.0281604.g001

This study also focuses on the sub-robot’s role that supports the main robot. The sub-robot can apologize after the main robot, although it can also fulfill a different role, such as cleaning up spilled food while the main robot apologizes for knocking it over. Investigating the effectiveness of the sub-robot’s behavior will also provide useful knowledge in collaboration with other robots in the context of mitigating failures. Based on these considerations, we pose another research question:

  1. Research question 2: Should the sub-robot only apologize, clean up, or do both?

Related works

Apology in human-robot interaction

Similar to apologies in human-human interaction, apologies from robots are useful to diminish failures. Robotics researchers implemented different kinds of apology behaviors for social robots and investigated their effectiveness. A typical approach is to design a verbal apology and offer compensation to mitigate a failure [24, 28] because verbal behavior design is useful for such robots as non-humanoids and simply designed social robots that lack sufficient degree of freedoms.

Non-verbal behavior design is another typical approach to mitigate failures. Researchers focused on reaction behaviors when robots fail to show that the robots understand their own mistakes [25, 26]. Apologies with non-verbal behaviors are limited to social robots with adequate capabilities to express various motions, such as humanoid robots, but integrating both verbal and non-verbal behaviors is probably more effective for mitigating failures [27]. Another study reported the importance of showing remorse in a robot apology, including both verbal and non-verbal behaviors [29]. Such apology behaviors are effective not only in human-robot interaction but also in human-agent interaction contexts [3538].

Another study, which focused on the appearance of robots in apology situations [39], reported that humanoid service robots can recover more fully than non-humanoid robots from a service failure by giving a sincere apology. This result suggests that the perceived impressions toward robots influence whether people can forgive them.

Although these studies showed the effectiveness of their behavior designs in apology situations, as described above, they mainly focused on how to design acceptable apologies with only one robot. Unlike these past studies, we focused on the effects of using multiple robots for apology situations and evaluated their effectiveness.

Number effects of social robots

Using multiple robots for services has become a popular approach. For example, researchers developed a multiple-robots-based information-providing system in which two robots discuss specific content, proving that such a style effectively attracts people in real environments [32, 40]. Researchers argued that observations of conversational-style information-providing between robots are easier for people than direct interaction with a social robot. Other studies reported the effectiveness of using multiple robots in education contexts for children [31, 34] and conversational services for the elderly [41]. These past studies described the effects of the behavior and the impression changes of the people involved by interaction with multiple robots.

Although these studies showed the effectiveness of using multiple robots in various situations, they focused less on apology situations. In other words, the effectiveness of multiple robots in apology situations has not yet been investigated. Although a past study in human science focused on costly group apology effects, it did not compare the number of persons who apologized [18]. Our study provides novel knowledge about the effectiveness of using multiple robots and their behavior designs in apology situations.

Experiment I

Experiment I has two aims: 1) investigating and comparing the effectiveness of apologies from two robots with an apology from just one to answer research question 1; 2) gathering opinions from participants about whether the second robot should only apologize or do other tasks. The second aim led to designing Experiment 2 which answers research question 2.

Hypothesis and prediction

An apology is an essential action to mitigate failure and be forgiven. Past studies in human science literature showed that the cost of apologies is an important factor in accepting them [16, 17, 42, 43]. Although such cost perspectives are not directly involved, robotics researchers investigated the effectiveness of sincere apology behaviors for service robots that fail in daily environments [2428].

To create a more acceptable apology for a robot, we focused on the number of robots that apologize. Past studies showed the effectiveness of using multiple robots in service contexts [31, 32, 34, 40, 41]. Moreover, using multiple robots simply increases the perceived cost of apologizing compared to a single robot. However, these studies did not deal with apology situations. Therefore, the effects of apologies from multiple robots are still unknown. Based on these considerations, we hypothesized that multiple robot apologies would have an advantage compared to one robot apology. Following this hypothesis, we made the following prediction:

  1. Prediction 1: If multiple robots apologize for their failure, people will accept the apologies and have more positive impressions than one robot’s apology.

Visual stimuli and conditions

We made videos where a food-service robot drops a customer’s order and apologizes. We used these videos in a web survey because of the obvious difficulties with such enactment in an actual restaurant. We used Pepper, developed by Softbank Robotics, which has 20 degrees of freedom (DOFs), stands 121 cm tall, and weighs 29 kg. Each video’s resolution was 1286 x 762 pixels; the fps was 30, and the lengths were about 27 seconds.

In this experiment, we prepared one factor (number factor, one/two conditions) and took two videos (Fig 2). In both videos, two robots appear (the main robot serves the food, and the sub-robot is another restaurant worker). First, the sub-robot crosses from left to right to explain that there are two robots in the video stimuli. Next, the main robot delivers an ice cream cone from the left and says, "Here you are." Unfortunately, it drops the cone and says, "Sorry." The main robot apologizes and offers to serve new ice cream without charging for it, and tell that no payment was required toward the original order too. The main robot’s speech contents (apologizing and offering compensation), which are identical between conditions, are based on a past study that investigated appropriate mitigating strategies for robot failures [24].

thumbnail
Fig 2. Visual stimuli of Experiment I.

Snapshots with specific timings (t = seconds).

https://doi.org/10.1371/journal.pone.0281604.g002

In the one-apology condition, the main robot apologizes, and the sub-robot only appears in the beginning. In the two-apologies condition, the sub-robot enters from the right and apologizes in the middle of the main robot’s apologizing. Therefore, the difference between the conditions is whether the sub-robot apologizes following the main robot after the latter’s mistake. We did not prepare a no-apology condition because a past study reported the effectiveness of a robot’s apology, i.e., a one-apology condition is better than a no-apology condition [24]. Therefore, we set the baseline of our study to the one-apology condition.

We note that the main robot’s behaviors are completely the same between the two conditions because we used the same programs to control its motions. The detailed behaviors are described in Table 1.

thumbnail
Table 1. The timeline of both the main and sub-robot in the two-apologies conditions.

https://doi.org/10.1371/journal.pone.0281604.t001

Measurement

We prepared a question with three choices to investigate the preferred number of robots during an apology. The participants chose their preferred number of apologizing robots from three choices: one, two, and either one. Having an additional choice as an intermediate option allows us to investigate whether an apology by two robots is effective. If we prepared only two choices (i.e., one or two), the participants are forced to choose even if they feel there is no actual difference.

We also measured questionnaire items to investigate the perceived feelings from various viewpoints based on past studies on the effects of service-recovery strategies and trust (all items are described in the (S1 File)). For the five questionnaires, we employed and modified the following: the forgiveness scale [44], negative word-of-mouth (NWOM) [44], performance trust from the multi-dimensional measure of trust (consisting of reliable and competent subscales) [45], moral trust from the multi-dimensional measure of trust (consisting of ethical, transparent, and benevolent subscales) [45], and intention to use (ITU) [46]. Each questionnaire item of each scale was assessed using a one-to-seven response format, where 1 was the most negative and 7 was the most positive.

We investigated the forgiveness, NWOM, and ITU scales about a shop and the trust scales toward the robot clerk serving the ice cream. We also prepared a free-text form to gather opinions from participants about whether only the second robot should apologize or do another task. The participants freely gave their opinions.

Due to the need for such screening of participants in web surveys [47, 48], we prepared two dummy questions to confirm whether participants carefully watched the videos and instructions at the end of the questionnaires. The first question asks the participants to answer what was dropped out by the robot in the video with four-choice questions. We excluded the participants who wrongly answered. The second question is based on the instruction manipulation check [48]. We translated the text of the example instruction manipulation check (Fig 1 in [48]) to Japanese text and prepared three responsible format items. The text asks participants to skip answers to confirm whether they have read the instruction. Therefore, we also excluded the participants who answered these dummy questions.

Procedure

All the procedures were approved by the Advanced Telecommunication Research Review Boards (20-501-3). The consent form is displayed in the beginning of the first webpage. The participants who only agreed with it can participate in the survey. First, the participants read explanations of the experiment and how to evaluate each video, and then we verified that they could clearly hear the video’s audio. We also displayed text to support imagining a situation where multiple robots work in a cafeteria.

Next, they observed a video with the one-apology or two-apologies robots apologizing conditions and answered the questions. They again watched a video with two-apologies or one-apology conditions and answered the questions. After watching both videos, they answered the three-choice question and completed the free-text form. Finally, they answered the dummy questions to determine how carefully they watched the video and to verify the quality of their answers.

Participants

The experiment was conducted using the participant pools of a Japanese crowd-sourcing company (CrowdWorks, Inc., which has more than three million registered workers). 203 people (101 women, 101 men, and 1 who declined to specify gender) joined our experiment. The screening process winnowed that number to 168 valid participants: 81 women, 86 men, and 1 who declined to specify.

Results

Preferred number of apologizing robots

Fig 3 shows the number of each answer to the three-choice question. We conducted a chi-square test and identified a significant difference (x2(2) = 19.609, p<0.01). Multiple comparisons with Ryan’s tests [49] revealed that the participants significantly preferred two robots more than the other candidates (one < two, p < 0.01 and no difference < two, p < 0.01).

thumbnail
Fig 3. Preferecnce of the participants.

Results of three-choice questions about the preferred number.

https://doi.org/10.1371/journal.pone.0281604.g003

Questionnaire results

Fig 4 shows the questionnaire results’ averages and standard error (S.E.). We conducted a paired-t test for each result, and all showed significant differences (forgiveness: t(1,167) = 4.034, p < 0.001, d = 0.778, NWOM: t(1,167) = 2.830, p = 0.005, d = 0.563, the reliable subscale; t(1,167) = 3.630, p < 0.001, d = 0.728, the competent subscale: t(1,167) = 6.115, p < 0.001, d = 0.785, the ethical subscale: t(1,167) = 6.757, p < 0.001, d = 0.665, the transparent subscale: t(1,167) = 2.585, p = 0.011, d = 0.716, the benevolent subscale: t(1,167) = 6.714, p < 0.001, d = 0.652, ITU: t(1,167) = 4.201, p < 0.001, d = 0.826). Thus, prediction 1 was supported.

thumbnail
Fig 4. Average and S.E. of questionnaire results of Experiment I.

The two-apologies condition is better than the one-apology condition from all questionnaire items.

https://doi.org/10.1371/journal.pone.0281604.g004

Free-description analysis

We gathered 287 sentences from the free-description forms. A coder categorized all the scripts as speech contents, bowing motions, additional behavior, and others (e.g., "no comments"). 37 sentences suggested adding speech content about the customers in the speech contents category, such as "Did anything get spilled on your clothes?" The remaining 45 sentences are just impressions of the contents and the tones of the robot’s voices. In the bowing category, 62 sentences suggested bowing more deeply, 39 suggested bowing longer, and nine suggested more bowing. In the additional behavior category, 40 sentences suggested wiping up the spill, 12 suggested that the robots perform separate roles, 10 suggested summoning a human clerk, and six suggested a replacement item instead of an apology. The other category had 27 sentences, although they did not suggest specific behaviors. We modified the robot behaviors in Experiment II based on the above information. The details are described in the next section.

Discussion

Potential benefits of apologies from multiple robots

Past studies reported the importance and effectiveness of apology behaviors of service robots [2229]. However, these studies mainly focused on an apology from one robot, i.e., less focused on the effects of apologies from multiple robots, although using multiple robots could influence peoples’ perceptions strongly more than using one robot [3034].

Our experiment results showed that the participants preferred apologies from two robots over one robot. In addition, all the questionnaire results showed they had positive impressions of two apologies: one from each robot. Higher forgiveness/ITU and lower NWOM scores indicate the robot’s apologies effectively mitigated the impressions of dissatisfaction caused by the robot’s mistake. Moreover, using multiple robots for apologies increased the trust scores of the robots and contributed to building a positive impression of the shop where the robots work. In particular, apologies from multiple robots were effectively perceived as more competent/ethical/benevolent perspectives from the viewpoint of the effect size. Therefore, these results answered research question 1: people preferred apologies from two robots more than from just one.

Even though multiple robots are effective for making an apology, preparing a sub-robot that only apologizes is inefficient. Each robot should have different roles, e.g., carrying a plate or taking orders at a restaurant. This result suggests that apologizing in support of another robot is a useful task for such robots.

Additional apology behavior design

Although our experiment results supported prediction 1, the participants provided several suggestions on their free-description forms and possible actions for the sub-robot: cleaning up the mess and performing clearly separate roles. Investigating an appropriate behavior design for the sub-robot might increase the possible benefits of having multiple robots in daily situations.

Our participants also provided suggestions to modify the apologizing behaviors, such as speech content and bowing behaviors. We modified the robot’s behaviors in Experiment II based on them. Note that we rejected the behavior that summons a human clerk because her behaviors and characteristics would influence the apologies.

Experiment II

Experiment II investigated the effects of the sub-robot and its behaviors, which were designed based on the free descriptions gathered from Experiment I.

Hypothesis and prediction

The results of Experiment I showed the effectiveness of apologies from two robots over one robot and suggested the importance of investigating different role effects for the sub-robot. The participant opinions from the free-description forms include designing a sub-robot that cleans the spill and performs a separate role. As suggested by the participants, such an additional role for the sub-robot might effectively mitigate the failure of the main robot. On the other hand, we must determine whether the sub-robot should apologize with the main robot before cleaning up or should it only clean up the spill. We hypothesized that the sub-robot’s additional action would improve perceived impressions. Based on this hypothesis, we made the following prediction:

  1. Prediction 2: If the sub-robot apologizes and cleans up during the main robot’s apology, people will accept the apologies and have more positive impressions than the sub-robot that only performs one action.

Visual stimuli and conditions

Based on the results from Experiment I, we prepared three different videos. Two robots appeared in all the videos, but the sub-robot’s behaviors are different depending on the conditions. The main robot’s apologizing behaviors are based on the results of Experiment I: deeper and longer bows and verbally apologizing to its customers.

The first condition is apologize-only (Fig 5a). Here, similar to the two apologies condition in Experiment I, only the sub-robot apologizes. The second condition is cleaning-up-only (Fig 5b); after the main robot’s apology, the sub-robot enters from the right with a broom and a dustpan. In other words, we clearly separated the robot roles into apologizing and cleaning up. The third condition combines both into a mixed condition (Fig 5c); the sub-robot first apologizes, as in the apologize-only condition, moves to the right, and returns with a broom and a dustpan. Its timings in the apologize-only and cleaning up-only conditions are different because they are based on the mixed condition’s timing. The details of the timings of behaviors are described in Table 2.

thumbnail
Fig 5. Three conditions.

Differences among them are sub-robot’s role (t = seconds). (a) Apologize-only condition: sub-robot apologizes. (b) Cleaning up-only condition: sub-robot brings a broom and dustpan. (c) Mixed condition: sub-robot apologizes and returns with a broom and dustpan.

https://doi.org/10.1371/journal.pone.0281604.g005

Measurement

We measured the same questionnaire items to investigate the effects of the robot apologies from Experiment I: the forgiveness scale [44], NWOM [44], the performance trust from the multi-dimensional measure of trust [45], moral trust from the multi-dimensional measure of trust [45], and intention to use (ITU) [46]. We excluded the three-choice question and the free-description forms in this experiment. Similar to Experiment I, we investigated the forgiveness, NWOM, and ITU scales about a shop and the trust scales toward the robot clerk serving the ice cream.

Procedure

Similar to Experiment I, all the procedures were approved by the Advanced Telecommunication Research Review Boards (20-501-3). First, the participants read explanations of the experiment and how to evaluate each video, and then we verified that they could clearly hear the video’s audio. Next they observed one of the three candidates, completed the questionnaires, and watched all three videos. The order of the videos was counterbalanced. Finally, they answered dummy questions to determine how carefully they watched the video and to verify the quality of their answers because past research reported the need for such screening of participants in web surveys [47, 48].

Participants

The experiment was conducted using the participant pools of the same Japanese survey company from Experiment 1. No participants in Experiment 1 were involved in Experiment 2; 609 people (337 women, 266 men, and 6 who declined to specify gender) joined. The screening process winnowed that number to 430 valid participants: 237 women, 188 men, and 5 who declined to specify.

Results

Questionnaire results

Fig 6 shows the questionnaire results. We conducted a repeated ANOVA for each result, and only three scales showed significant differences: (forgiveness: F(2,858) = 5.536, p = 0.004, partial η2 = 0.013, the reliable subscale; F(2,858) = 3.401, p = 0.034, partial η2 = 0.008, the competence subscale: F(2, 858) = 5.096, p = 0.006, partial η2 = 0.012). Multiple comparison with the Bonferroni method for forgiveness showed significant differences between conditions: apology-only < mixed (p = 0.011) and cleaning-up-only < mixed (p = 0.023). Multiple comparisons with the Bonferroni method for the reliable subscale showed significant differences between conditions: apology-only < mixed (p = 0.033). Multiple comparison with it for the competent subscale showed significant differences between conditions: apology-only < mixed (p = 0.008) and cleaning-up-only < mixed (p = 0.040).

thumbnail
Fig 6. Average and S.E. of questionnaire results of Experiment II.

The mixed condition is better than apology and cleaning-up conditions in forgiveness and competent scales.

https://doi.org/10.1371/journal.pone.0281604.g006

Other scales did not show significant differences (NWOM: F(2,858) = 1.143, p = 0.319, partial η2 = 0.003, the ethical subscale: F(2,858) = 1.792, p = 0.167, partial η2 = 0.004, the transparent subscale: F(2,858) = 1.931, p = 0.146, partial η2 = 0.004, the benevolent subscale: F(2,858) = 0.814, p = 0.444, partial η2 = 0.002, ITU: F(2,858) = 2.875, p = 0.057, partial η2 = 0.007). Thus, prediction 2 was partially supported.

Discussion

Potential benefits of apologizing in multiple roles

Our first experiment showed that the effectiveness of the apologies from multiple robots was preferred more than one robot, but the behavior design of the additional robot was relatively simple. We thought that investigating different behavioral effects of the additional robots would contribute to mitigating failures, then we compared three different behaviors in this experiment.

Our experiment results showed that participants preferred the mixed condition in the context of forgiveness, two of the trust viewpoints (reliable and competent) compared to the apology condition, and in the context of the competent trust viewpoint more than the cleaning-up-only condition. We believe that the differences in the competent scores are appropriate because multiple roles contributed to efficiently solving the problem. Thus, although the mixed condition did not show significant differences in the other measurements, apologies from two robots and separating their roles after they apologize is an appropriate strategy to mitigate the consequences of a robot’s mistakes. These results showed that having multiple roles enables the sub-robot to mitigate another robot’s mistake more efficiently. Similar to the results of Experiment 1, having multiple apologizing roles has merit for the shop where the robots are working.

The cleaning-up-only condition did not show any significant differences compared to the apology condition, although it did show some disadvantages compared to the mixed condition. Although the participants in Experiment 1 suggested cleaning up and separating the roles of the robots, they were ineffective unless combined with an apology. In the mixed condition, the participants had slightly more sincere and efficient impressions toward the robots compared to the other conditions where the sub-robot only did one task. we thought that the weak effects of cleaning-up-only conditions are mainly two kinds. The first one is the modified behaviors of the main robot; most of the questionnaire results of the apologize-only condition in Experiment 2 are better than the results of the two condition in Experiment I. The second one is that the suggested behaviors from the participants in Experiment 1 are less effective, although many of them thought such behaviors would be appropriate.

General discussion

Since completely eliminating mistakes by service robots in actual environments is nearly impossible, effectively mitigating strategies is important. This study provided several implications about such strategies for robot mistakes. We thought that an apology from multiple robots might increase the perceived costs related to acceptance. Although quantifying the literal cost of increasing the number of robots is difficult, our experiment results showed that an apology from multiple robots is more acceptable. Here, we discuss implications and limitations based on two experiment results.

Implications for apologies from multiple robots

Why are two robots better than one robot in an apology situation? One possible reason is that the sub-robot’s apology resembles a helping behavior toward the main robot, not only an apologizing behavior toward the customer. Past studies reported that helping or generosity is driven by receiving or observing assistance [50, 51]. If the participants thought that the sub-robot was both apologizing and supporting the main robot, such help-observation effects might have influenced their perceived impressions. This viewpoint seems related to increasing trust scores; people may more highly evaluate a robot supported by another robot than a robot without any support.

Another possible reason is a perceived apology cost. Simply using two robots incurs a higher cost than just using one. Related to this point, the participant might perceive the apology to be sincerer when two robots apologize. A past human-robot apology study also reported the effectiveness of a robot’s sincere apology [2125]. These studies expressed sincerity by a single robot’s behaviors, and using multiple robots should increase the perception of such sincere impressions. In addition, we should consider a peer pressure effect from multiple robots. People’s decision-making is influenced by the opinions of others [52, 53]. Several past studies reported conformity effects from robot groups, which change people’s behaviors and impressions [54, 55]. Although the number of robots in our study is just two (not many robots), the participants might have felt peer pressure from the robots.

Based on these considerations, one remaining unknown point is the most effective number of robots. Perhaps simply increasing the number of robots is not effective for acceptable apology behavior because people might become suspicious and cynical toward so many apologizing robots. Since past studies with multiple robots for information-providing situations focused less on an effective number of social robots, this issue remains unresolved. Another study that focused on peer pressure effects from social robots investigated the effective number and reported that six robots strengthened the peer pressure effects more than two or four robots [55], although apologies from six robots is excessive. In fact, a few participants in Experiment I described that a scenario where more robots apologized might create negative impressions. Investigating the best balance between the number of apologizing robots and the perceived cost is interesting future work.

Another implication is the robot’s appearance. In this study, we used Pepper as a humanoid robot that can serve items. However, if the robot’s appearance is more humanlike, people’s perceptions will be different. Using different robots also influences forgiveness; if the sub-robot is more human-like or has more authority, people will accept its apology more. On the other hand, if the sub-robot seems to have less authority, people’s perceptions will also be different. Using android robots with a human-like appearance [56, 57] might provide additional knowledge for robot-apology behavior designs.

Implications for apologies by robots and people

Until robot clerks are sufficiently diffused into human society, human and robot clerks might work together. Although this study did not investigate the effects of apologies due to metaphysical differences, we believe that a human clerk’s apology will be more effective than one from a robot clerk. A past study reported the importance of apology cost toward acceptance [31], and human clerks currently have higher social status than robot clerks.

We believe that investigating the effects of additional apologies by robots with human mistakes will be fruitful, e.g., a human clerk’s mistake, and then a robot clerk apologizing with the human clerk. In addition, the effectiveness of a robot’s apology will be influenced by relationships and such roles as a child/childcare robot, senior citizen/nursing robot, student/teacher robot, etc.

Limitations

This study has several limitations, including using specific robots, situations, and context. Therefore, as described above, using different robots, environments, and contexts, such as the responsibility of failure is not lying to the robot clerk (e.g., an accident caused by other people), would have different results with multiple robots’ apologies. It investigated the perceived impressions toward a robot’s failure by videos, not actual failures in daily situations. In fact, related studies in behavioral economics are performed with real money (typically in the trust game, e.g., [58, 59]). Therefore, conducting additional studies in actual settings is critical to investigate the effects of multiple robot apologies. On the other hand, conducting real situations under such experiment settings is complicated by ethical viewpoints. Therefore, several related studies in human-robot interaction conducted a video-based survey to investigate the effects of robot behaviors in specific situations with difficulties that reproduced real situations [25, 60]. We used a video-based survey approach to provide insight into designing a robot’s behaviors in specific situations, including mitigating failures by robots.

Since we only conducted our experiments with Japanese participants, we must consider generality and cultural differences [15, 61, 62]. One future work will investigate the effectiveness of using multiple robots to mitigate failures in different cultures. Related to such future work, investigating the influence of different kinds of robots and behaviors, e.g., more humanlike robots, is essential for understanding the relationships between cultural differences and acceptable apologies by robots. In addition, different environments such as high-end restaurants will undoubtedly also influence the perceived impressions of robot failures.

Conclusion

This study investigated the effects of apologies by multiple robots to mitigate the mistakes made by robots. We first prepared different visual stimuli where one or two robots apologized in a restaurant environment. The main robot makes a mistake and apologizes, and then the sub-robot apologizes with the main robot. We conducted a web-survey-based experiment to investigate the perceived impressions whose results showed that apologies from two robots are preferred over an apology from just one. We also gathered suggestions from participants about modifying the sub-robot’s behaviors while the main robot apologizes. Based on such free descriptions, we created three different behaviors for the sub-robot: apologize-only, cleaning-up-only, and mixed combinations. We conducted another web-survey-based experiment to investigate the perceived impressions and identified a part of the advantages of both combinations over apologizing-only and cleaning-up-only. Our results suggest the importance of supporting a robot that makes a mistake with another robot that can apologize with it, and provide the evidences about the effectiveness of using multiple robot apologies.

Supporting information

S1 File. Anonymized data set and questionnaire items.

https://doi.org/10.1371/journal.pone.0281604.s001

(XLSX)

References

  1. 1. Jiang Y., and Wen J., “Effects of COVID-19 on hotel marketing and management: a perspective article,” International Journal of Contemporary Hospitality Management, 2020.
  2. 2. Kim S. S., Kim J., Badu-Baiden F., Giroux M., and Choi Y., “Preference for robot service or human service in hotels? Impacts of the COVID-19 pandemic,” International Journal of Hospitality Management, vol. 93, pp. 102795, 2021.
  3. 3. Niemelä M., Heikkilä P., Lammi H., and Oksman V., "A social robot in a shopping mall: studies on acceptance and stakeholder expectations," Social Robots: Technological, Societal and Ethical Aspects of Human-Robot Interaction, pp. 119–144: Springer, 2019.
  4. 4. Brengman M., De Gauquier L., Willems K., and Vanderborght B., “From stopping to shopping: An observational study comparing a humanoid service robot with a tablet service kiosk to attract and convert shoppers,” Journal of Business Research, vol. 134, pp. 263–274, 2021.
  5. 5. Kanda T., Shiomi M., Miyashita Z., Ishiguro H., and Hagita N., “A communication robot in a shopping mall,” Robotics, IEEE Transactions on, vol. 26, no. 5, pp. 897–913, 2010.
  6. 6. P. Kong, H. Cornet, and F. Frenkler, “Personas and emotional design for public service robots: A case study with autonomous vehicles in public transportation,” in 2018 international conference on cyberworlds (cw), pp. 284–287, 2018.
  7. 7. M. Tonkin, J. Vitale, S. Herse, M.-A. Williams, W. Judge, and X. Wang, “Design methodology for the ux of hri: A field study of a commercial social robot at an airport,” in Proceedings of the 2018 ACM/IEEE International Conference on Human-Robot Interaction, pp. 407–415, 2018.
  8. 8. Shiomi M., Sakamoto D., Kanda T., Ishi C. T., Ishiguro H., and Hagita N., “Field Trial of a Networked Robot at a Train Station,” International Journal of Social Robotics, vol. 3, no. 1, pp. 27–40, 2010.
  9. 9. S. Okano, N. Matsuhira, E. Shimokawara, T. Yamaguchi, and M. Narita, “Employing robots in a museum environment: Design and implementation of collaborative robot network,” in 2019 16th International Conference on Ubiquitous Robots (UR), pp. 224–227, 2019.
  10. 10. S. Pliasa, A. M. Velentza, A. G. Dimitriou, and N. Fachantidis, “Interaction of a Social Robot with Visitors inside a Museum through RFID Technology,” in 2021 6th International Conference on Smart and Sustainable Technologies (SpliTech), pp. 01–06, 2021.
  11. 11. Shiomi M., Kanda T., Ishiguro H., and Hagita N., “Interactive Humanoid Robots for a Science Museum,” IEEE Intelligent Systems, no. 2, pp. 25–32, 2007.
  12. 12. Do H. M., Pham M., Sheng W., Yang D., and Liu M., “RiSH: A robot-integrated smart home for elderly care,” Robotics and Autonomous Systems, vol. 101, pp. 74–92, 2018.
  13. 13. Portugal D., Alvito P., Christodoulou E., Samaras G., and Dias J., “A study on the deployment of a service robot in an elderly care center,” International Journal of Social Robotics, vol. 11, no. 2, pp. 317–341, 2019.
  14. 14. Sumioka H., Yamato N., Shiomi M., and Ishiguro H., “A minimal design of a human infant presence: a case study toward interactive doll therapy for older adults with dementia,” Frontiers in Robotics and AI, pp. 164, 2021.
  15. 15. Wagatsuma H., and Rosett A., “The implications of apology: Law and culture in Japan and the United States,” Law & Soc’y Rev., vol. 20, pp. 461, 1986.
  16. 16. Abeler J., Calaki J., Andree K., and Basek C., “The power of apology,” Economics Letters, vol. 107, no. 2, pp. 233–235, 2010.
  17. 17. Martinez-Vaquero L. A., Han T. A., Pereira L. M., and Lenaerts T., “Apology and forgiveness evolve to resolve failures in cooperative agreements,” Scientific reports, vol. 5, no. 1, pp. 1–12, 2015. pmid:26057819
  18. 18. Ohtsubo Y., Matsunaga M., Himichi T., Suzuki K., Shibata E., Hori R., et al. “Costly group apology communicates a group’s sincere “intention”,” Social Neuroscience, vol. 15, no. 2, pp. 244–254, 2020.
  19. 19. T. A. Han, L. M. Pereira, F. C. Santos, and T. Lenaerts, “Why is it so hard to say sorry? evolution of apology with commitments in the iterated Prisoner’s Dilemma,” in Proceedings of the Twenty-Third international joint conference on Artificial Intelligence, pp. 177–183, 2013.
  20. 20. T. Han, L. Pereira, F. Santos, and T. Lenaerts, “Why is it so hard to say sorry,” in BNAIC 2013: Proceedings of the 25th Benelux Conference on Artificial Intelligence, Delft, The Netherlands, November 7–8, pp. 4–9, 2013.
  21. 21. A. Mahmood, J. W. Fung, I. Won, and C.-M. Huang, “Owning Mistakes Sincerely: Strategies for Mitigating AI Errors,” in Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems, New Orleans, LA, USA, pp. Article 578, 2022.
  22. 22. Honig S., and Oron-Gilad T., “Understanding and resolving failures in human-robot interaction: Literature review and model development,” Frontiers in psychology, vol. 9, pp. 861, 2018. pmid:29962981
  23. 23. P. Gieselmann, “Comparing error-handling strategies in human-human and human-robot dialogues,” in Proc. 8th Conf. Nat. Language Process.(KONVENS). Konstanz, Germany, pp. 24–31, 2006.
  24. 24. M. K. Lee, S. Kiesler, J. Forlizzi, S. Srinivasa, and P. Rybski, “Gracefully mitigating breakdowns in robotic services,” in Human-Robot Interaction (HRI), 2010 5th ACM/IEEE International Conference on, pp. 203–210, 2010.
  25. 25. L. Takayama, D. Dooley, and W. Ju, “Expressing thought: improving robot readability with animation principles,” in Proceedings of the 6th international conference on Human-robot interaction, pp. 69–76, 2011.
  26. 26. Shiomi M., Nakagawa K., and Hagita N., “Design of a gaze behavior at a small mistake moment for a robot,” Interaction Studies, vol. 14, no. 3, pp. 317–328, 2013.
  27. 27. A. Hamacher, N. Bianchi-Berthouze, A. G. Pipe, and K. Eder, “Believing in BERT: Using expressive communication to enhance trust and counteract operational error in physical Human-robot interaction,” in 2016 25th IEEE international symposium on robot and human interactive communication (RO-MAN), pp. 493–500, 2016.
  28. 28. Cameron D., de Saille S., Collins E. C., Aitken J. M., Cheung H., Chua A., et al. “The effect of social-cognitive recovery strategies on likability, capability and trust in social robots,” Computers in Human Behavior, vol. 114, pp. 106561, 2021.
  29. 29. B. L. Pompe, E. Velner, and K. P. Truong, “The Robot That Showed Remorse: Repairing Trust with a Genuine Apology,” in 2022 31st IEEE International Conference on Robot and Human Interactive Communication (RO-MAN), pp. 260–265, 2022.
  30. 30. Shiomi M., Okumura S., Kimoto M., Iio T., and Shimohara K., “Two is better than one: Social rewards from two agents enhance offline improvements in motor skills more than single agent,” PloS one, vol. 15, no. 11, pp. e0240622, 2020. pmid:33147230
  31. 31. Shiomi M., Tamura Y., Kimoto M., Iio T., Akahane-Yamada R., and Shimohara K., “Two is better than one: verification of the effect of praise from two robots on pre-school children’s learning time,” Advanced Robotics, vol. 35, no. 19, pp. 1132–1144, 2021.
  32. 32. Sakamoto D., Hayashi K., Kanda T., Shiomi M., Koizumi S., Ishiguro H., et al. “Humanoid Robots as a Broadcasting Communication Medium in Open Public Spaces,” International Journal of Social Robotics, vol. 1, no. 2, pp. 157–169, 2009.
  33. 33. Iio T., Yoshikawa Y., and Ishiguro H., “Double-meaning agreements by two robots to conceal incoherent agreements to user’s opinions,” Advanced Robotics, vol. 35, no. 19, pp. 1145–1155, 2021.
  34. 34. Tamura Y., Shiomi M., Kimoto M., Iio T., Shimohara K., and Hagita N., “Robots as an interactive-social medium in storytelling to multiple children,” Interaction Studies, vol. 22, no. 1, pp. 110–140, 2021.
  35. 35. Z. Ashktorab, M. Jain, Q. V. Liao, and J. D. Weisz, “Resilient chatbots: Repair strategy preferences for conversational breakdowns,” in Proceedings of the 2019 CHI conference on human factors in computing systems, pp. 1–12, 2019.
  36. 36. C.-H. Li, S.-F. Yeh, T.-J. Chang, M.-H. Tsai, K. Chen, and Y.-J. Chang, “A conversation analysis of non-progress and coping strategies with a banking task-oriented chatbot,” in Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, pp. 1–12, 2020.
  37. 37. Yuasa M., “Do You Forgive Past Mistakes of Animated Agents? A Study of Instances of Assistance by Animated Agents,” Journal of Advanced Computational Intelligence and Intelligent Informatics, vol. 24, no. 3, pp. 404–412, 2020.
  38. 38. A. Mahmood, J. W. Fung, I. Won, and C.-M. Huang, “Owning Mistakes Sincerely: Strategies for Mitigating AI Errors,” in CHI Conference on Human Factors in Computing Systems, pp. 1–11, 2022.
  39. 39. Choi S., Mattila A. S., and Bolton L. E., “To err is human (-oid): how do consumers react to robot service failure and recovery?,” Journal of Service Research, vol. 24, no. 3, pp. 354–371, 2021.
  40. 40. Iio T., Yoshikawa Y., and Ishiguro H., “Retaining Human-Robots Conversation: Comparing Single Robot to Multiple Robots in a Real Event,” Journal of Advanced Computational Intelligence Intelligent Informatics, vol. 21, no. 4, pp. 675–685, 2017.
  41. 41. Nishio T., Yoshikawa Y., Iio T., Chiba M., Asami T., Isoda Y., et al. “Actively listening twin robots for long-duration conversation with the elderly,” ROBOMECH Journal, vol. 8, no. 1, pp. 1–10, 2021.
  42. 42. Wei C., Liu M. W., and Keh H. T., “The road to consumer forgiveness is paved with money or apology? The roles of empathy and power in service recovery,” Journal of Business Research, vol. 118, pp. 321–334, 2020.
  43. 43. Halperin B., Ho B., List J. A., and Muir I., “Toward an understanding of the economics of apologies: evidence from a large-scale natural field experiment,” The Economic Journal, vol. 132, no. 641, pp. 273–298, 2022.
  44. 44. Casidy R., and Shin H., “The effects of harm directions and service recovery strategies on customer forgiveness and negative word-of-mouth intentions,” Journal of Retailing and Consumer Services, vol. 27, pp. 103–112, 2015.
  45. 45. Malle B. F., and Ullman D., “A multi-dimensional conception and measure of human-robot trust,” Trust in Human-Robot Interaction, pp. 3, 2020.
  46. 46. Heerink M., Ben K., Evers V., and Wielinga B., “The influence of social presence on acceptance of a companion robot by older people,” Journal of Physical Agents, vol. 2, no. 2, pp. 33–40, 2008.
  47. 47. J. S. Downs, M. B. Holbrook, S. Sheng, and L. F. Cranor, “Are your participants gaming the system? screening mechanical turk workers,” in Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, Atlanta, Georgia, USA, pp. 2399–2402, 2010.
  48. 48. Oppenheimer D. M., Meyvis T., and Davidenko N., “Instructional manipulation checks: Detecting satisficing to increase statistical power,” Journal of experimental social psychology, vol. 45, no. 4, pp. 867–872, 2009.
  49. 49. Ryan T. H., “Significance tests for multiple comparison of proportions, variances, and other statistics,” Psychological bulletin, vol. 57, no. 4, pp. 318, 1960. pmid:14440422
  50. 50. Fowler J. H., and Christakis N. A., “Cooperative behavior cascades in human social networks,” Proceedings of the National Academy of Sciences, vol. 107, no. 12, pp. 5334–5338, 2010. pmid:20212120
  51. 51. Tsvetkova M., and Macy M. W., “The Social Contagion of Generosity,” PLOS ONE, vol. 9, no. 2, pp. e87275, 2014. pmid:24551053
  52. 52. Asch S. E., “Opinions and Social Pressure,” Scientific American, vol. 193, pp. 17–26, 1955.
  53. 53. Asch S. E., “Effects of group pressure upon the modification and distortion of judgments,” Groups, leadership, and men., pp. 222–236, 1951.
  54. 54. Vollmer A.-L., Read R., Trippas D., and Belpaeme T., “Children conform, adults resist: A robot group induced peer pressure on normative social conformity,” Science Robotics, vol. 3, no. 21, 2018. pmid:33141720
  55. 55. Shiomi M., and Hagita N., “Do the number of robots and the participant’s gender influence conformity effect from multiple robots?,” Advanced Robotics, vol. 33, no. 15–16, pp. 756–763, 2019.
  56. 56. D. F. Glas, T. Minato, C. T. Ishi, T. Kawahara, and H. Ishiguro, “Erica: The erato intelligent conversational android,” in Robot and Human Interactive Communication (RO-MAN), 2016 25th IEEE International Symposium on, pp. 22–29, 2016.
  57. 57. M. Shiomi, H. Sumioka, K. Sakai, T. Funayama, and T. Minato, “SŌTO: An Android Platform with a Masculine Appearance for Social Touch Interaction,” in Companion of the 2020 ACM/IEEE International Conference on Human-Robot Interaction, Cambridge, United Kingdom, pp. 447–449, 2020.
  58. 58. Schniter E., Sheremeta R. M., and Sznycer D., “Building and rebuilding trust with promises and apologies,” Journal of Economic Behavior & Organization, vol. 94, pp. 242–256, 2013.
  59. 59. Schniter E., Shields T. W., and Sznycer D., “Trust in humans and robots: Economically similar but emotionally different,” Journal of Economic Psychology, vol. 78, pp. 102253, 2020.
  60. 60. E. K. Duarte, M. Shiomi, A. Vinel, and M. Cooney, “Robot Self-defense: Robots Can Use Force on Human Attackers to Defend Victims*,” in 2022 31st IEEE International Conference on Robot and Human Interactive Communication (RO-MAN), pp. 1606–1613, 2022.
  61. 61. Guan X., Park H. S., and Lee H. E., “Cross-cultural differences in apology,” International Journal of Intercultural Relations, vol. 33, no. 1, pp. 32–45, 2009.
  62. 62. Maddux W. W., Kim P. H., Okumura T., and Brett J. M., “Cultural differences in the function and meaning of apologies,” International negotiation, vol. 16, no. 3, pp. 405–425, 2011.