YaBeSH Engineering and Technology Library

    • Journals
    • PaperQuest
    • YSE Standards
    • YaBeSH
    • Login
    View Item 
    •   YE&T Library
    • ASME
    • Journal of Autonomous Vehicles and Systems
    • View Item
    •   YE&T Library
    • ASME
    • Journal of Autonomous Vehicles and Systems
    • View Item
    • All Fields
    • Source Title
    • Year
    • Publisher
    • Title
    • Subject
    • Author
    • DOI
    • ISBN
    Advanced Search
    JavaScript is disabled for your browser. Some features of this site may not work without it.

    Archive

    SafetyAware Adversarial Inverse Reinforcement Learning for Highway Autonomous Driving

    Source: Journal of Autonomous Vehicles and Systems:;2022:;volume( 001 ):;issue: 004::page 41004
    Author:
    Li, Fangjian;Wagner, John;Wang, Yue
    DOI: 10.1115/1.4053427
    Publisher: The American Society of Mechanical Engineers (ASME)
    Abstract: Inverse reinforcement learning (IRL) has been successfully applied in many robotics and autonomous driving studies without the need for handtuning a reward function. However, it suffers from safety issues. Compared to the reinforcement learning algorithms, IRL is even more vulnerable to unsafe situations as it can only infer the importance of safety based on expert demonstrations. In this paper, we propose a safetyaware adversarial inverse reinforcement learning (SAIRL) algorithm. First, the control barrier function is used to guide the training of a safety critic, which leverages the knowledge of system dynamics in the sampling process without training an additional guiding policy. The trained safety critic is then integrated into the discriminator to help discern the generated data and expert demonstrations from the standpoint of safety. Finally, to further enforce the importance of safety, a regulator is introduced in the loss function of the discriminator training to prevent the recovered reward function from assigning high rewards to the risky behaviors. We tested our SAIRL in the highway autonomous driving scenario. Comparing to the original AIRL algorithm, with the same level of imitation learning performance, the proposed SAIRL can reduce the collision rate by 32.6%.
    • Download: (1.031Mb)
    • Show Full MetaData Hide Full MetaData
    • Get RIS
    • Item Order
    • Go To Publisher
    • Price: 5000 Rial
    • Statistics

      SafetyAware Adversarial Inverse Reinforcement Learning for Highway Autonomous Driving

    URI
    http://yetl.yabesh.ir/yetl1/handle/yetl/4288679
    Collections
    • Journal of Autonomous Vehicles and Systems

    Show full item record

    contributor authorLi, Fangjian;Wagner, John;Wang, Yue
    date accessioned2023-04-06T12:52:43Z
    date available2023-04-06T12:52:43Z
    date copyright2/4/2022 12:00:00 AM
    date issued2022
    identifier otherjavs_1_4_041004.pdf
    identifier urihttp://yetl.yabesh.ir/yetl1/handle/yetl/4288679
    description abstractInverse reinforcement learning (IRL) has been successfully applied in many robotics and autonomous driving studies without the need for handtuning a reward function. However, it suffers from safety issues. Compared to the reinforcement learning algorithms, IRL is even more vulnerable to unsafe situations as it can only infer the importance of safety based on expert demonstrations. In this paper, we propose a safetyaware adversarial inverse reinforcement learning (SAIRL) algorithm. First, the control barrier function is used to guide the training of a safety critic, which leverages the knowledge of system dynamics in the sampling process without training an additional guiding policy. The trained safety critic is then integrated into the discriminator to help discern the generated data and expert demonstrations from the standpoint of safety. Finally, to further enforce the importance of safety, a regulator is introduced in the loss function of the discriminator training to prevent the recovered reward function from assigning high rewards to the risky behaviors. We tested our SAIRL in the highway autonomous driving scenario. Comparing to the original AIRL algorithm, with the same level of imitation learning performance, the proposed SAIRL can reduce the collision rate by 32.6%.
    publisherThe American Society of Mechanical Engineers (ASME)
    titleSafetyAware Adversarial Inverse Reinforcement Learning for Highway Autonomous Driving
    typeJournal Paper
    journal volume1
    journal issue4
    journal titleJournal of Autonomous Vehicles and Systems
    identifier doi10.1115/1.4053427
    journal fristpage41004
    journal lastpage4100413
    page13
    treeJournal of Autonomous Vehicles and Systems:;2022:;volume( 001 ):;issue: 004
    contenttypeFulltext
    DSpace software copyright © 2002-2015  DuraSpace
    نرم افزار کتابخانه دیجیتال "دی اسپیس" فارسی شده توسط یابش برای کتابخانه های ایرانی | تماس با یابش
    yabeshDSpacePersian
     
    DSpace software copyright © 2002-2015  DuraSpace
    نرم افزار کتابخانه دیجیتال "دی اسپیس" فارسی شده توسط یابش برای کتابخانه های ایرانی | تماس با یابش
    yabeshDSpacePersian