ISSN: 2090-9888
Lovotics
Make the best use of Scientific Research and information from our 700+ peer reviewed, Open Access Journals that operates with the help of 50,000+ Editorial Board Members and esteemed reviewers and 1000+ Scientific associations in Medical, Clinical, Pharmaceutical, Engineering, Technology and Management Fields.
Meet Inspiring Speakers and Experts at our 3000+ Global Conferenceseries Events with over 600+ Conferences, 1200+ Symposiums and 1200+ Workshops on Medical, Pharma, Engineering, Science, Technology and Business

Designing Humanoid Robots with Novel Roles and Social Abilities

Maya Dimitrova1,2* and Hiroaki Wagatsuma2,3

1Institute of Systems Engineering and Robotics, Bulgarian Academy of Sciences (BAS), Akad, Bulgaria

2Graduate School of Life Science and Systems Engineering, Kyushu Institute of Technology, Japan

3RIKEN BSI, 2-1, Hirosawa, Wako-shi, Saitama, 351-0198, Japan

*Corresponding Author:
Maya Dimitrova
Institute of Systems Engineering and
Robotics, Bulgarian Academy of Sciences
(BAS), Akad, G. Bonchev, Str., Bl. 2
POB 79, 1113 Sofia, Bulgaria
Tel: +359-882-866-270
E-mail: dimitrova@iser.bas.bg

Received Date:January 30, 2015; Accepted Date: January 31, 2015; Published Date: February 10, 2015

Citation: Dimitrova M, Wagatsuma H (2015) Designing Humanoid Robots with Novel Roles and Social Abilities. Lovotics 3:112. doi:10.4172/2090-9888.1000112

Copyright: © 2015 Dimitrova M, et al. This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.

Visit for more related articles at Lovotics

Abstract

This mini-review describes aspects of human-robot interaction to be taken into account when designing humanoid robots with novel roles and social abilities. The review accentuates the psychological complexity that is necessary to be made inherent in the design of humanoid robotic technology. Some recent studies of robot acceptance are summarized leading to the proposal for more complex synthetic sensors being needed in novel humanoid robots. The perspective is designing based on modeling attitude (the social level of human robot interaction), but not opinion (psychological level), which can be a valuable aim for humanoid robotics.

Keywords

Humanoid robots; Synthetic sensors; Uncanny valley; Psychology; Behavior; Attitude; Modeling

Introduction

The interest in humanoid robotics is constantly increasing and is motivating novel solutions for personal assistance, elderly care, child education or even for “designing robots for well-being” [1]. In building humanoid robots with novel roles in our society more social and psychological factors will be of major concern in addition to the safety and acceptability issues of having robots performing close to the human and engaging them in joint tasks [2-4]. The aim of the present mini-review is to outline some possible novel roles of humanoid robots accentuating on the psychological complexity that is necessary to be made inherent in the design of humanoid robotic technology.

An especially designed study has revealed that, unlike the common view “that robots should do dangerous, dirty or dull jobs; public opinion favors the view that robots should perform tasks requiring memorization, keen perceptual skills and service orientation. People are preferred for occupations that require artistry, evaluation, judgment and diplomacy” [5]. The factor levels of the study were: jobs “performed either by robots or people” vs. “performed by both robots and people” and what people think robots “could do” or “should do”. The Participants in the experiment, who were not expert with robots, filled in a web-based survey with questions like the following: “If, hypothetically, robots and people could be fire inspectors, who would you prefer?” with possible answers: “Robots either\or people” or Robots\both\people”. The study demonstrated that people accepted with little surprise the possibility to be asked questions about the utility of having humans or robots performing as professionals. Therefore new roles for humanoid robots can be formulated, like being teachers, driving instructors, medical care advisors or counselors, even friends and fashion experts. In this future social medium, densely populated with humanoid robots with different professional roles, what type of human-robot sameness, closeness, or “bond” has to be established, is one of the crucial questions to be asked, as stated in [2].

Synthetic Sensor Design for Improved Human-Robot ��?Mutual��? Understanding

One aspect of the area of synthetic sensor design for humanoid robots will have to deal with much more complex sensations being function of integrative ‘brain-like’ analyzers, such as a “gaze sensor” as it is proposed in [6]. A gaze sensor is a complex algorithm, providing robots with understanding of the attention attracted to others or their communicative intentions similarly to the way people perceive them at some subtle level of awareness. The term ‘affection’ is becoming more and more characteristic of the process of natural and intuitive humanrobot interaction, when the anthropomorphism is the underlying notion of this process according to [7], calling for the design of an ‘affection sensor’, i.e. for modeling of complex multidimensional aspects of the psychological side of the human-robot interaction process. Designing social sensors as integrative ’brain-like’ analyzers is a next step in the development of humanoid robots with novel roles, which can be possibly comprised of infinite number of dimension combinations of features underlying the process of communicating with a humanoid robot. For example, methods for defining the characteristics of a human communicating with an intelligent computer interface, for adapting to personality dimensions like generosity, sociability and cautiousness/ risk taking, i.e. in the design of the relevant synthetic sensors are described in [8]. Examples from experimental studies have provided evidence that the similarity of humans and robots is defined by people in terms of deep, essential, rather than shallow, characteristic features, according to the classical taxonomy formulated by Rosch to explain concept formation based on the ‘family resemblance’ principle [9]. Actually, people (not being experts with robot design) do not implicitly assume the android a robot – on the contrary – the mechanical outlook is usually combined with human-like intelligence or sensitivity - at the level of human intuitive assumptions. This is confirmed by a study on “reported characteristics for home computers, robots and human beings” [10]. In the condition Human-Computer, participants were asked to list features of humans and subsequently to indicate which features were shared with computers. The labels Human-Robot, Robot- Human, and Computer-Human denote the respective other conditions in the study. In the Human-Computer condition the proportion of shared features between a computer and a human is 0.06, in Human- Robot - 0.42, in Robot-Human - 0.31 and in Computer-Human - 0.11. Users tend to report more features shared between humans and robots than with other intelligent devices like computers, on the one hand. On the other – this is rather characteristic, than essential – i.e. below 50% of the reported features. Even more interesting is the result on attributing facial characteristics to robots, humans and computers. Given that only 8% of robot features are facial features, whereas the comparable percentage for human beings is about 30%, robots are “not stereotypically defined by their face” (italics by C. Ramey) [10]. At the same time, face features are often reported in a form like: “Scary face”, “Imitation of human face”, “Pair of eyes” or “Rectangle mouth”. Therefore, based on the deep, essential features shared between humans and robots, it is logical to assume that a humanoid robot has to be able to display empathy – as one of the deepest humane features guiding our social life [11], to keep memories of our autobiographical experiences [12] and be able to anticipate future events based on one-time experiences [13]. We can imagine iFoxy, a humanoid robot like the Fox in the book of Antoine de Saint Exupéry, saying to the Little Prince: “It is only with the heart that one can see rightly; what is essential is invisible to the eye” [14] in a therapeutic session to reduce the anxiety of a human patient.

Overcoming the ��?Visceral Uneasiness��? of Human-Robot Interaction

Recent studies reveal that people perceive simultaneously multiple aspects of the agency attributed to the robot like, for example, “visceral factors of interaction”, “social mechanics” and “social structures” [15]. The authors relate the “visceral factors” to the ‘uncanny valley’ phenomenon, defined first by Mori [16] – whenever the surface, physical attributes of the robot exceed a certain degree of resemblance to the human – feelings of unpleasantness and fear emerge in people communicating with the robots. One of the hypotheses explaining it is that on a subtle discriminative, i.e. ‘visceral’, level, sensing of ‘strangeness’ of robot behavior emerges provoking negative reaction and thus presenting an obstacle to the flawless human-robot communication [16-18].

A novel hypothesis of the categorical nature of the ‘uncanny valley’ phenomenon has been recently proposed by Moore [18]. The advantage of his model is the mathematical description of the observed discrepancy between the subjective comfort of the interaction with the robot and the sudden repulsion by the realization that the creature we are communicating with is a non-human. His model represents ‘the human’ as a normal distribution of ‘objects’ possessing features, defining the human as a conceptual category. This distribution is characterized by its mean, standard deviation and mathematical function that is descriptive of the form of the distribution and delimits the category boundaries. Representatives of the category “human”, sharing very typical or essential for the category features, in the abstract perceptual (i.e. internal) space are nearer the center of the distribution, whereas the representatives with less typical (i.e. surface or characteristic) features are at the distribution outskirts – i.e. near the category boundaries. Inside the distribution, the probability of occurrence of a ‘target’ category – a human with typical ‘human’ features is higher; therefore there is better ‘predictability’ (as subjective anticipation) to encounter a typical human ‘target’.

This distribution is combined by Moore with the distribution of the category ‘non-human’, which has bigger variance, hence broader span below the distribution function. The mean of the second - “background” - distribution does not typically coincide with the mean of the ‘target’ distribution in terms of the amount of features, describing essentially a ‘human’ in a categorical sense. However, if these coinciding features make the distribution means close enough, at some moment the forms of the overlapping functions, denoting category boundaries, form a ‘function’ with two optimums in Mori’s sense – one positive when a humanoid robot resembles a stuffed animal - and one negative – when the interaction with the robot generates the feeling of communicating with a zombie. This form of the function was plotted in 3D by Duffy with dimensions - acceptance, efficiency and emotion. Acceptance and efficiency are of positive sign, but emotion follows Mori’s function separating two 3D ‘hills’ – of the so called “economic ideal” and the “ultimate ideal”, the latter representing the “perfect artificial human” in Duffy’s words [19].

The model of Moore plots ‘affinity’ in Mori’s sense as familiarity plus ‘perceptual tension’. The subtracting of the perceptual tension, playing the role of internal weighting factor, from the familiarity, will predict the observed phenomenon well. “If the weighting factor is small or zero, then the implication is that the observer does not notice (or does not care), if perceptual cues are in conflict. If the weighting factor is large, then it indicates a strong sensitivity to differential cues on the part of an observer. The weighting is thus a key property of an observer, not of a stimulus” [19]. The model of the ‘visceral’ uneasiness of the human-robot interaction process, proposed by Moore, aims at explaining a variety of psychological phenomena when perceiving conflicting cues in an observed scene can invoke repulsion, anger or aggression and as such has deep societal validity. As guiding principle for humanoid robot design for fulfilling novel professional roles, based on the current mini-review, the following conceptualization of the process of human-robot interaction is proposed: People perceive other people (or humanoid robots) simultaneously at three levels of abstraction - physical, social and psychological - and form anticipation of the actions of the others depending on the respective level. The physical level is the case of ‘predicting behavior in response to behavior’ – as physical objects clashing or avoiding each other. The social level is similar to the “social mechanics” proposal of [15] – when people play teams with people/robots and display attitudes of tolerance and affection towards people/robots. We call this case ‘predicting behavior in response to attitude’. This social level of communication is when people perceive the attitudes of the agent – human or robot – positive, empathic, friendly, assistive, negative, rude, hostile - which current robotics is completely capable of reinstating and where most effort is to be devoted in humanoid robot design.

The psychological level of ‘predicting behavior in response to opinion’, in our view, is the “uncanny” case. Whenever people react as if they feel that the behavior of the robot is guided not just by attitude (social level), but by opinion (psychological level), by some kind of awareness like the one produced by a ‘gaze sensor’, we expect to observe the ‘uncanny valley’ phenomenon. Robots need synthetic sensors like the ‘gaze sensor’ but they need not reinstate situations where the human ‘gaze sensor’ is on. They can rather reinstate feelings of positive attitude, friendliness, trust and compassion. Special questionnaires, distinguishing feelings close to perception of attitude from perception of opinion in human-robot interaction need to be designed to explore the validity of this hypothesis.

Conclusion

Humanoid robotics is gradually taking novel roles and is performing tasks usually referred to by people as “professional”. This will require the design of novel synthetic sensors, capable of capturing complex psychological phenomena in a human-like manner, as the present mini- review has revealed. We propose to keep in mind the tendency of people to animate creatures (cartoons, animation, fairy tales and fantasy) as part of their internal confidence with the external living and nonliving world. This confidence is the result of concentrating primarily on the social level of interaction with inanimate creatures. The perspective is designing based on the social level of human-robot interaction process or on modeling attitude - friendly, supportive and empathic - which can be a valuable aim for humanoid robotics.

Acknowledgements

This work was performed while the first author was on a FY2014 JSPS Invitation Fellowship Program (Short Term) for Research in Japan ID No. S-14156 at Kyushu Institute of Technology (KYUTECH).

References

--
Post your comment

Share This Article

Recommended Journals

Article Usage

  • Total views: 16767
  • [From(publication date):
    March-2015 - Aug 16, 2024]
  • Breakdown by view type
  • HTML page views : 15724
  • PDF downloads : 1043
Top