Cyber-Human Approach For Learning Human Intention And Shape Robotic Behavior Based On Task Demonstration

Recent developments in artificial intelligence enabled training of autonomous robots without human supervision. Even without human supervision during training, current models have yet to be human-engineered and have neither guarantees to match human expectation nor perform within safety bounds. This...

Full description

Saved in:
Bibliographic Details
Published in:2018 International Joint Conference on Neural Networks (IJCNN) pp. 1 - 7
Main Authors: Goecks, Vinicius G., Gremillion, Gregory M., Lehman, Hannah C., Nothwang, William D.
Format: Conference Proceeding
Language:English
Published: IEEE 01-07-2018
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Abstract Recent developments in artificial intelligence enabled training of autonomous robots without human supervision. Even without human supervision during training, current models have yet to be human-engineered and have neither guarantees to match human expectation nor perform within safety bounds. This paper proposes CyberSteer to leverage human-robot interaction and align goals between humans and robotic intelligent agents. Based on human demonstration of the task, CyberSteer learns an intrinsic reward function used by the human demonstrator to pursue the goal of the task. The learned intrinsic human function shapes the robotic behavior during training through deep reinforcement learning algorithms, removing the need for environment-dependent or hand-engineered reward signal. Two different hypotheses were tested, both using non-expert human operators for initial demonstration of a given task or desired behavior: one training a deep neural network to classify human-like behavior and other training a behavior cloning deep neural network to suggest actions. In this experiment, CyberSteer was tested in a high-fidelity unmanned air system simulation environment, Microsoft AirSim. The simulated aerial robot performed collision avoidance through a clustered forest environment using forward-looking depth sensing. The performance of CyberSteer is compared to behavior cloning algorithms and reinforcement learning algorithms guided by handcrafted reward functions. Results show that the human-learned intrinsic reward function can shape the behavior of robotic systems and have better task performance guiding reinforcement learning algorithms compared to standard human-handcrafted reward functions.
AbstractList Recent developments in artificial intelligence enabled training of autonomous robots without human supervision. Even without human supervision during training, current models have yet to be human-engineered and have neither guarantees to match human expectation nor perform within safety bounds. This paper proposes CyberSteer to leverage human-robot interaction and align goals between humans and robotic intelligent agents. Based on human demonstration of the task, CyberSteer learns an intrinsic reward function used by the human demonstrator to pursue the goal of the task. The learned intrinsic human function shapes the robotic behavior during training through deep reinforcement learning algorithms, removing the need for environment-dependent or hand-engineered reward signal. Two different hypotheses were tested, both using non-expert human operators for initial demonstration of a given task or desired behavior: one training a deep neural network to classify human-like behavior and other training a behavior cloning deep neural network to suggest actions. In this experiment, CyberSteer was tested in a high-fidelity unmanned air system simulation environment, Microsoft AirSim. The simulated aerial robot performed collision avoidance through a clustered forest environment using forward-looking depth sensing. The performance of CyberSteer is compared to behavior cloning algorithms and reinforcement learning algorithms guided by handcrafted reward functions. Results show that the human-learned intrinsic reward function can shape the behavior of robotic systems and have better task performance guiding reinforcement learning algorithms compared to standard human-handcrafted reward functions.
Author Goecks, Vinicius G.
Gremillion, Gregory M.
Lehman, Hannah C.
Nothwang, William D.
Author_xml – sequence: 1
  givenname: Vinicius G.
  surname: Goecks
  fullname: Goecks, Vinicius G.
  organization: Texas A&M University, College Station, USA
– sequence: 2
  givenname: Gregory M.
  surname: Gremillion
  fullname: Gremillion, Gregory M.
  organization: US Army Research Laboratory, Adelphi, USA
– sequence: 3
  givenname: Hannah C.
  surname: Lehman
  fullname: Lehman, Hannah C.
  organization: Texas A&M University, College Station, USA
– sequence: 4
  givenname: William D.
  surname: Nothwang
  fullname: Nothwang, William D.
  organization: US Army Research Laboratory, Adelphi, USA
BookMark eNotkM1OwkAUhUejiYC8gG7mBYrz25lZQhXBEEgU1-TO9NZWZdq01YS3FwOrs_jOdxZnSK5iHZGQO84mnDP3sHzJ1uuJYNxOrLJOO31Bxs5YrpljKeMqvSQDwVOeKMXMDRl23SdjQjonB6TMDh7bZPGzh0inTdPWEEo6r1u6QmhjFT_oiS1jj7Gv6mMr5vSthAbpa-3rvgp0hiX8VkdnBh3mdBPpFrov-oj7OnZ9C__aLbku4LvD8TlH5H3-tM0WyWrzvMymq6TiRvcJmLyQWkoPTHklAqjUeq7QhFyLwgSFgSnLlWXaCRFUUIVzXkMBxjM0XI7I_Wm3QsRd01Z7aA-78zHyD0qBWXA
ContentType Conference Proceeding
DBID 6IE
6IH
CBEJK
RIE
RIO
DOI 10.1109/IJCNN.2018.8489595
DatabaseName IEEE Electronic Library (IEL) Conference Proceedings
IEEE Proceedings Order Plan (POP) 1998-present by volume
IEEE Xplore All Conference Proceedings
IEEE Electronic Library Online
IEEE Proceedings Order Plans (POP) 1998-present
DatabaseTitleList
Database_xml – sequence: 1
  dbid: RIE
  name: IEEE Electronic Library Online
  url: http://ieeexplore.ieee.org/Xplore/DynWel.jsp
  sourceTypes: Publisher
DeliveryMethod fulltext_linktorsrc
Discipline Computer Science
EISBN 9781509060146
1509060146
EISSN 2161-4407
EndPage 7
ExternalDocumentID 8489595
Genre orig-research
GroupedDBID 29I
29O
6IE
6IF
6IH
6IK
6IL
6IM
6IN
AAJGR
ABLEC
ADZIZ
ALMA_UNASSIGNED_HOLDINGS
BEFXN
BFFAM
BGNUA
BKEBE
BPEOZ
CBEJK
CHZPO
IEGSK
IJVOP
IPLJI
M43
OCL
RIE
RIL
RIO
RNS
ID FETCH-LOGICAL-i175t-a7df3533ba04b42ca468b14e7cd52f7c4ec04814805922c4c4f99b5afa7b0e713
IEDL.DBID RIE
IngestDate Wed Jun 26 19:31:36 EDT 2024
IsPeerReviewed false
IsScholarly false
Language English
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-i175t-a7df3533ba04b42ca468b14e7cd52f7c4ec04814805922c4c4f99b5afa7b0e713
PageCount 7
ParticipantIDs ieee_primary_8489595
PublicationCentury 2000
PublicationDate 2018-July
PublicationDateYYYYMMDD 2018-07-01
PublicationDate_xml – month: 07
  year: 2018
  text: 2018-July
PublicationDecade 2010
PublicationTitle 2018 International Joint Conference on Neural Networks (IJCNN)
PublicationTitleAbbrev IJCNN
PublicationYear 2018
Publisher IEEE
Publisher_xml – name: IEEE
SSID ssj0023993
ssj0002685453
Score 1.7343118
Snippet Recent developments in artificial intelligence enabled training of autonomous robots without human supervision. Even without human supervision during training,...
SourceID ieee
SourceType Publisher
StartPage 1
SubjectTerms Cloning
Human-robot interaction
Intelligent robots
Neural networks
Robot learning
Robot sensing systems
Semisupervised learning
Shape
Task analysis
Training
unmanned air systems
Title Cyber-Human Approach For Learning Human Intention And Shape Robotic Behavior Based On Task Demonstration
URI https://ieeexplore.ieee.org/document/8489595
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://sdu.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwlV09T8MwELVoJ6YCLeJbNzCS1nGc2h5LPwQMBdEisVWOPyhCJFXaDvx77CQNQmJhixIpinzn3MvlvXcIXVtNucFcB5FnuFIbRoELs8tlTCTG2kouvFD4bsamr3w09jY5N7UWxhhTkM9M1x8W__J1pra-VdbjlItYxA3UYIKXWq26n0L63IGB-i3sJZvRTiSDRe_-YTideiYX71Z3-TVOpagmk9b_nuMAdX5kefBUF5xDtGfSI9TazWWAapu20XL4lZg8KPrzMKhMw2GS5VC5qb5Bea3gr_vIwCDVMFvKlYHnLMlcMkFlnJjDratzGh5TmMv1B4zMp0eUZd500MtkPB_eBdVEheDdwYRNIJm2kQN4icQ0oURJ2udJSA1TOiaWKWqU94-h3IEuQhRV1AqRxNJKlmDjvmePUTPNUnOCoM-wViFWJKSWSoYl57YvZawEYd4G7xS1_cItVqVpxqJas7O_T5-jfR-bkgd7gZqbfGsuUWOtt1dFmL8BybWnRw
link.rule.ids 310,311,782,786,791,792,798,27934,54767
linkProvider IEEE
linkToHtml http://sdu.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwlV1NT4NAEN1oPeipamv8dg8epV2WhV2OtR9ptaKxNfHWLPthjREa2h789-4CxZh48UYgIWRnlnkM770B4FpLwhRi0vEsw5Vo13NMmE0uI8wRkpqz0AqFhxMavbJe39rk3FRaGKVUTj5TLXuY_8uXqVjbVlmbERb6ob8NdnxCA1qotaqOCg6YgQPVe9iKNr2NTAaF7dFdN4osl4u1yvv8GqiS15NB_X9Psg-aP8I8-FSVnAOwpZJDUN9MZoDlRm2AefcrVpmTd-hhp7QNh4M0g6Wf6hssruUMdhsb2EkknMz5QsHnNE5NOsHSOjGDt6bSSfiYwClffsCe-rSYssicJngZ9KfdoVPOVHDeDVBYOZxK7RmIF3NEYoIFJwGLXaKokD7WVBAlrIMMYQZ2YSyIIDoMY59rTmOkzBftEaglaaKOAQwoksJFArtEE04RZ0wHnPsixNQa4Z2Ahl242aKwzZiVa3b69-krsDucPoxn41F0fwb2bJwKVuw5qK2ytboA20u5vsxD_g3Gb6qY
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=2018+International+Joint+Conference+on+Neural+Networks+%28IJCNN%29&rft.atitle=Cyber-Human+Approach+For+Learning+Human+Intention+And+Shape+Robotic+Behavior+Based+On+Task+Demonstration&rft.au=Goecks%2C+Vinicius+G.&rft.au=Gremillion%2C+Gregory+M.&rft.au=Lehman%2C+Hannah+C.&rft.au=Nothwang%2C+William+D.&rft.date=2018-07-01&rft.pub=IEEE&rft.eissn=2161-4407&rft.spage=1&rft.epage=7&rft_id=info:doi/10.1109%2FIJCNN.2018.8489595&rft.externalDocID=8489595