YaBeSH Engineering and Technology Library

    • Journals
    • PaperQuest
    • YSE Standards
    • YaBeSH
    • Login
    View Item 
    •   YE&T Library
    • ASCE
    • Journal of Infrastructure Systems
    • View Item
    •   YE&T Library
    • ASCE
    • Journal of Infrastructure Systems
    • View Item
    • All Fields
    • Source Title
    • Year
    • Publisher
    • Title
    • Subject
    • Author
    • DOI
    • ISBN
    Advanced Search
    JavaScript is disabled for your browser. Some features of this site may not work without it.

    Archive

    A Reinforcement Learning Method for Multiasset Roadway Improvement Scheduling Considering Traffic Impacts

    Source: Journal of Infrastructure Systems:;2022:;Volume ( 028 ):;issue: 004::page 04022033
    Author:
    Weiwen Zhou
    ,
    Elise Miller-Hooks
    ,
    Kostas G. Papakonstantinou
    ,
    Shelley Stoffels
    ,
    Sue McNeil
    DOI: 10.1061/(ASCE)IS.1943-555X.0000702
    Publisher: ASCE
    Abstract: Maintaining roadway pavements and bridge decks is key to providing high levels of service for road users. However, improvement actions incur downtime. These actions are typically scheduled by asset class, yet implemented on any asset type, they have network-wide impacts on traffic performance. This paper presents a bilevel program wherein the upper level involves a Markov decision process (MDP) through which potential roadway improvement actions across asset classes are prioritized and scheduled. The MDP approach considers uncertainty in component deterioration effects, while incorporating the benefits of implemented improvement actions. The upper level takes as input traffic flow estimates obtained from a lower-level user equilibrium traffic formulation that recognizes changes in capacities determined by decisions taken in the upper level. Because an exact solution of this bilevel, stochastic, dynamic program is formidable, a deep reinforcement learning (DRL) method is developed. The model and solution methodology were tested on a hypothetical problem from the literature. The importance of obtaining optimal activity plans that account for downtime effects, traffic congestion impacts, uncertainty in deterioration processes, and multiasset classes is demonstrated.
    • Download: (3.779Mb)
    • Show Full MetaData Hide Full MetaData
    • Get RIS
    • Item Order
    • Go To Publisher
    • Price: 5000 Rial
    • Statistics

      A Reinforcement Learning Method for Multiasset Roadway Improvement Scheduling Considering Traffic Impacts

    URI
    http://yetl.yabesh.ir/yetl1/handle/yetl/4287740
    Collections
    • Journal of Infrastructure Systems

    Show full item record

    contributor authorWeiwen Zhou
    contributor authorElise Miller-Hooks
    contributor authorKostas G. Papakonstantinou
    contributor authorShelley Stoffels
    contributor authorSue McNeil
    date accessioned2022-12-27T20:39:33Z
    date available2022-12-27T20:39:33Z
    date issued2022/12/01
    identifier other(ASCE)IS.1943-555X.0000702.pdf
    identifier urihttp://yetl.yabesh.ir/yetl1/handle/yetl/4287740
    description abstractMaintaining roadway pavements and bridge decks is key to providing high levels of service for road users. However, improvement actions incur downtime. These actions are typically scheduled by asset class, yet implemented on any asset type, they have network-wide impacts on traffic performance. This paper presents a bilevel program wherein the upper level involves a Markov decision process (MDP) through which potential roadway improvement actions across asset classes are prioritized and scheduled. The MDP approach considers uncertainty in component deterioration effects, while incorporating the benefits of implemented improvement actions. The upper level takes as input traffic flow estimates obtained from a lower-level user equilibrium traffic formulation that recognizes changes in capacities determined by decisions taken in the upper level. Because an exact solution of this bilevel, stochastic, dynamic program is formidable, a deep reinforcement learning (DRL) method is developed. The model and solution methodology were tested on a hypothetical problem from the literature. The importance of obtaining optimal activity plans that account for downtime effects, traffic congestion impacts, uncertainty in deterioration processes, and multiasset classes is demonstrated.
    publisherASCE
    titleA Reinforcement Learning Method for Multiasset Roadway Improvement Scheduling Considering Traffic Impacts
    typeJournal Article
    journal volume28
    journal issue4
    journal titleJournal of Infrastructure Systems
    identifier doi10.1061/(ASCE)IS.1943-555X.0000702
    journal fristpage04022033
    journal lastpage04022033_15
    page15
    treeJournal of Infrastructure Systems:;2022:;Volume ( 028 ):;issue: 004
    contenttypeFulltext
    DSpace software copyright © 2002-2015  DuraSpace
    نرم افزار کتابخانه دیجیتال "دی اسپیس" فارسی شده توسط یابش برای کتابخانه های ایرانی | تماس با یابش
    yabeshDSpacePersian
     
    DSpace software copyright © 2002-2015  DuraSpace
    نرم افزار کتابخانه دیجیتال "دی اسپیس" فارسی شده توسط یابش برای کتابخانه های ایرانی | تماس با یابش
    yabeshDSpacePersian