Cyber-Human Approach For Learning Human Intention And Shape Robotic Behavior Based On Task Demonstration
Recent developments in artificial intelligence enabled training of autonomous robots without human supervision. Even without human supervision during training, current models have yet to be human-engineered and have neither guarantees to match human expectation nor perform within safety bounds. This...
Saved in:
Published in: | 2018 International Joint Conference on Neural Networks (IJCNN) pp. 1 - 7 |
---|---|
Main Authors: | , , , |
Format: | Conference Proceeding |
Language: | English |
Published: |
IEEE
01-07-2018
|
Subjects: | |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Abstract | Recent developments in artificial intelligence enabled training of autonomous robots without human supervision. Even without human supervision during training, current models have yet to be human-engineered and have neither guarantees to match human expectation nor perform within safety bounds. This paper proposes CyberSteer to leverage human-robot interaction and align goals between humans and robotic intelligent agents. Based on human demonstration of the task, CyberSteer learns an intrinsic reward function used by the human demonstrator to pursue the goal of the task. The learned intrinsic human function shapes the robotic behavior during training through deep reinforcement learning algorithms, removing the need for environment-dependent or hand-engineered reward signal. Two different hypotheses were tested, both using non-expert human operators for initial demonstration of a given task or desired behavior: one training a deep neural network to classify human-like behavior and other training a behavior cloning deep neural network to suggest actions. In this experiment, CyberSteer was tested in a high-fidelity unmanned air system simulation environment, Microsoft AirSim. The simulated aerial robot performed collision avoidance through a clustered forest environment using forward-looking depth sensing. The performance of CyberSteer is compared to behavior cloning algorithms and reinforcement learning algorithms guided by handcrafted reward functions. Results show that the human-learned intrinsic reward function can shape the behavior of robotic systems and have better task performance guiding reinforcement learning algorithms compared to standard human-handcrafted reward functions. |
---|---|
AbstractList | Recent developments in artificial intelligence enabled training of autonomous robots without human supervision. Even without human supervision during training, current models have yet to be human-engineered and have neither guarantees to match human expectation nor perform within safety bounds. This paper proposes CyberSteer to leverage human-robot interaction and align goals between humans and robotic intelligent agents. Based on human demonstration of the task, CyberSteer learns an intrinsic reward function used by the human demonstrator to pursue the goal of the task. The learned intrinsic human function shapes the robotic behavior during training through deep reinforcement learning algorithms, removing the need for environment-dependent or hand-engineered reward signal. Two different hypotheses were tested, both using non-expert human operators for initial demonstration of a given task or desired behavior: one training a deep neural network to classify human-like behavior and other training a behavior cloning deep neural network to suggest actions. In this experiment, CyberSteer was tested in a high-fidelity unmanned air system simulation environment, Microsoft AirSim. The simulated aerial robot performed collision avoidance through a clustered forest environment using forward-looking depth sensing. The performance of CyberSteer is compared to behavior cloning algorithms and reinforcement learning algorithms guided by handcrafted reward functions. Results show that the human-learned intrinsic reward function can shape the behavior of robotic systems and have better task performance guiding reinforcement learning algorithms compared to standard human-handcrafted reward functions. |
Author | Goecks, Vinicius G. Gremillion, Gregory M. Lehman, Hannah C. Nothwang, William D. |
Author_xml | – sequence: 1 givenname: Vinicius G. surname: Goecks fullname: Goecks, Vinicius G. organization: Texas A&M University, College Station, USA – sequence: 2 givenname: Gregory M. surname: Gremillion fullname: Gremillion, Gregory M. organization: US Army Research Laboratory, Adelphi, USA – sequence: 3 givenname: Hannah C. surname: Lehman fullname: Lehman, Hannah C. organization: Texas A&M University, College Station, USA – sequence: 4 givenname: William D. surname: Nothwang fullname: Nothwang, William D. organization: US Army Research Laboratory, Adelphi, USA |
BookMark | eNotkM1OwkAUhUejiYC8gG7mBYrz25lZQhXBEEgU1-TO9NZWZdq01YS3FwOrs_jOdxZnSK5iHZGQO84mnDP3sHzJ1uuJYNxOrLJOO31Bxs5YrpljKeMqvSQDwVOeKMXMDRl23SdjQjonB6TMDh7bZPGzh0inTdPWEEo6r1u6QmhjFT_oiS1jj7Gv6mMr5vSthAbpa-3rvgp0hiX8VkdnBh3mdBPpFrov-oj7OnZ9C__aLbku4LvD8TlH5H3-tM0WyWrzvMymq6TiRvcJmLyQWkoPTHklAqjUeq7QhFyLwgSFgSnLlWXaCRFUUIVzXkMBxjM0XI7I_Wm3QsRd01Z7aA-78zHyD0qBWXA |
ContentType | Conference Proceeding |
DBID | 6IE 6IH CBEJK RIE RIO |
DOI | 10.1109/IJCNN.2018.8489595 |
DatabaseName | IEEE Electronic Library (IEL) Conference Proceedings IEEE Proceedings Order Plan (POP) 1998-present by volume IEEE Xplore All Conference Proceedings IEEE Electronic Library Online IEEE Proceedings Order Plans (POP) 1998-present |
DatabaseTitleList | |
Database_xml | – sequence: 1 dbid: RIE name: IEEE Electronic Library Online url: http://ieeexplore.ieee.org/Xplore/DynWel.jsp sourceTypes: Publisher |
DeliveryMethod | fulltext_linktorsrc |
Discipline | Computer Science |
EISBN | 9781509060146 1509060146 |
EISSN | 2161-4407 |
EndPage | 7 |
ExternalDocumentID | 8489595 |
Genre | orig-research |
GroupedDBID | 29I 29O 6IE 6IF 6IH 6IK 6IL 6IM 6IN AAJGR ABLEC ADZIZ ALMA_UNASSIGNED_HOLDINGS BEFXN BFFAM BGNUA BKEBE BPEOZ CBEJK CHZPO IEGSK IJVOP IPLJI M43 OCL RIE RIL RIO RNS |
ID | FETCH-LOGICAL-i175t-a7df3533ba04b42ca468b14e7cd52f7c4ec04814805922c4c4f99b5afa7b0e713 |
IEDL.DBID | RIE |
IngestDate | Wed Jun 26 19:31:36 EDT 2024 |
IsPeerReviewed | false |
IsScholarly | false |
Language | English |
LinkModel | DirectLink |
MergedId | FETCHMERGED-LOGICAL-i175t-a7df3533ba04b42ca468b14e7cd52f7c4ec04814805922c4c4f99b5afa7b0e713 |
PageCount | 7 |
ParticipantIDs | ieee_primary_8489595 |
PublicationCentury | 2000 |
PublicationDate | 2018-July |
PublicationDateYYYYMMDD | 2018-07-01 |
PublicationDate_xml | – month: 07 year: 2018 text: 2018-July |
PublicationDecade | 2010 |
PublicationTitle | 2018 International Joint Conference on Neural Networks (IJCNN) |
PublicationTitleAbbrev | IJCNN |
PublicationYear | 2018 |
Publisher | IEEE |
Publisher_xml | – name: IEEE |
SSID | ssj0023993 ssj0002685453 |
Score | 1.7343118 |
Snippet | Recent developments in artificial intelligence enabled training of autonomous robots without human supervision. Even without human supervision during training,... |
SourceID | ieee |
SourceType | Publisher |
StartPage | 1 |
SubjectTerms | Cloning Human-robot interaction Intelligent robots Neural networks Robot learning Robot sensing systems Semisupervised learning Shape Task analysis Training unmanned air systems |
Title | Cyber-Human Approach For Learning Human Intention And Shape Robotic Behavior Based On Task Demonstration |
URI | https://ieeexplore.ieee.org/document/8489595 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://sdu.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwlV09T8MwELVoJ6YCLeJbNzCS1nGc2h5LPwQMBdEisVWOPyhCJFXaDvx77CQNQmJhixIpinzn3MvlvXcIXVtNucFcB5FnuFIbRoELs8tlTCTG2kouvFD4bsamr3w09jY5N7UWxhhTkM9M1x8W__J1pra-VdbjlItYxA3UYIKXWq26n0L63IGB-i3sJZvRTiSDRe_-YTideiYX71Z3-TVOpagmk9b_nuMAdX5kefBUF5xDtGfSI9TazWWAapu20XL4lZg8KPrzMKhMw2GS5VC5qb5Bea3gr_vIwCDVMFvKlYHnLMlcMkFlnJjDratzGh5TmMv1B4zMp0eUZd500MtkPB_eBdVEheDdwYRNIJm2kQN4icQ0oURJ2udJSA1TOiaWKWqU94-h3IEuQhRV1AqRxNJKlmDjvmePUTPNUnOCoM-wViFWJKSWSoYl57YvZawEYd4G7xS1_cItVqVpxqJas7O_T5-jfR-bkgd7gZqbfGsuUWOtt1dFmL8BybWnRw |
link.rule.ids | 310,311,782,786,791,792,798,27934,54767 |
linkProvider | IEEE |
linkToHtml | http://sdu.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwlV1NT4NAEN1oPeipamv8dg8epV2WhV2OtR9ptaKxNfHWLPthjREa2h789-4CxZh48UYgIWRnlnkM770B4FpLwhRi0vEsw5Vo13NMmE0uI8wRkpqz0AqFhxMavbJe39rk3FRaGKVUTj5TLXuY_8uXqVjbVlmbERb6ob8NdnxCA1qotaqOCg6YgQPVe9iKNr2NTAaF7dFdN4osl4u1yvv8GqiS15NB_X9Psg-aP8I8-FSVnAOwpZJDUN9MZoDlRm2AefcrVpmTd-hhp7QNh4M0g6Wf6hssruUMdhsb2EkknMz5QsHnNE5NOsHSOjGDt6bSSfiYwClffsCe-rSYssicJngZ9KfdoVPOVHDeDVBYOZxK7RmIF3NEYoIFJwGLXaKokD7WVBAlrIMMYQZ2YSyIIDoMY59rTmOkzBftEaglaaKOAQwoksJFArtEE04RZ0wHnPsixNQa4Z2Ahl242aKwzZiVa3b69-krsDucPoxn41F0fwb2bJwKVuw5qK2ytboA20u5vsxD_g3Gb6qY |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=2018+International+Joint+Conference+on+Neural+Networks+%28IJCNN%29&rft.atitle=Cyber-Human+Approach+For+Learning+Human+Intention+And+Shape+Robotic+Behavior+Based+On+Task+Demonstration&rft.au=Goecks%2C+Vinicius+G.&rft.au=Gremillion%2C+Gregory+M.&rft.au=Lehman%2C+Hannah+C.&rft.au=Nothwang%2C+William+D.&rft.date=2018-07-01&rft.pub=IEEE&rft.eissn=2161-4407&rft.spage=1&rft.epage=7&rft_id=info:doi/10.1109%2FIJCNN.2018.8489595&rft.externalDocID=8489595 |