YaBeSH Engineering and Technology Library

    • Journals
    • PaperQuest
    • YSE Standards
    • YaBeSH
    • Login
    View Item 
    •   YE&T Library
    • ASCE
    • Journal of Water Resources Planning and Management
    • View Item
    •   YE&T Library
    • ASCE
    • Journal of Water Resources Planning and Management
    • View Item
    • All Fields
    • Source Title
    • Year
    • Publisher
    • Title
    • Subject
    • Author
    • DOI
    • ISBN
    Advanced Search
    JavaScript is disabled for your browser. Some features of this site may not work without it.

    Archive

    Real-Time Pump Scheduling in Water Distribution Networks Using Deep Reinforcement Learning

    Source: Journal of Water Resources Planning and Management:;2025:;Volume ( 151 ):;issue: 006::page 04025012-1
    Author:
    Shengwei Pei
    ,
    Lan Hoang
    ,
    Guangtao Fu
    ,
    David Butler
    DOI: 10.1061/JWRMD5.WRENG-6476
    Publisher: American Society of Civil Engineers
    Abstract: Pump scheduling in water distribution networks (WDNs) influences energy efficiency and water supply reliability. Conventional optimization methods usually face challenges in intensive computational requirements and water demand uncertainty handling. This study presents a deep reinforcement learning (DRL) method, i.e., proximal policy optimization (PPO), for real-time pump scheduling in WDNs. The PPO agents are trained to develop offline policies in advance, avoiding the online optimization process during the scheduling period. They are compared with genetic algorithm-based baseline methods, including online optimization methods (i.e., scenario-specific optimization and model predictive control) and a robust optimization method, using the Anytown and D-town networks. The results obtained indicate that the PPO agents outperform the robust optimization method regarding operational cost and robustness to demand uncertainty and achieve the same level of pump scheduling performance as the online optimization methods. Including the demand and time information in the input for PPO agent training improves the performance of the DRL method. A smaller scheduling step size could improve the performance of PPO agents. This study illustrates the potential of PPO in real-time pump scheduling in WDNs and provides insight into the development and application of this method in practice.
    • Download: (1.832Mb)
    • Show Full MetaData Hide Full MetaData
    • Get RIS
    • Item Order
    • Go To Publisher
    • Price: 5000 Rial
    • Statistics

      Real-Time Pump Scheduling in Water Distribution Networks Using Deep Reinforcement Learning

    URI
    http://yetl.yabesh.ir/yetl1/handle/yetl/4309080
    Collections
    • Journal of Water Resources Planning and Management

    Show full item record

    contributor authorShengwei Pei
    contributor authorLan Hoang
    contributor authorGuangtao Fu
    contributor authorDavid Butler
    date accessioned2026-02-16T21:21:20Z
    date available2026-02-16T21:21:20Z
    date copyright2025/06/01
    date issued2025
    identifier otherJWRMD5.WRENG-6476.pdf
    identifier urihttp://yetl.yabesh.ir/yetl1/handle/yetl/4309080
    description abstractPump scheduling in water distribution networks (WDNs) influences energy efficiency and water supply reliability. Conventional optimization methods usually face challenges in intensive computational requirements and water demand uncertainty handling. This study presents a deep reinforcement learning (DRL) method, i.e., proximal policy optimization (PPO), for real-time pump scheduling in WDNs. The PPO agents are trained to develop offline policies in advance, avoiding the online optimization process during the scheduling period. They are compared with genetic algorithm-based baseline methods, including online optimization methods (i.e., scenario-specific optimization and model predictive control) and a robust optimization method, using the Anytown and D-town networks. The results obtained indicate that the PPO agents outperform the robust optimization method regarding operational cost and robustness to demand uncertainty and achieve the same level of pump scheduling performance as the online optimization methods. Including the demand and time information in the input for PPO agent training improves the performance of the DRL method. A smaller scheduling step size could improve the performance of PPO agents. This study illustrates the potential of PPO in real-time pump scheduling in WDNs and provides insight into the development and application of this method in practice.
    publisherAmerican Society of Civil Engineers
    titleReal-Time Pump Scheduling in Water Distribution Networks Using Deep Reinforcement Learning
    typeJournal Article
    journal volume151
    journal issue6
    journal titleJournal of Water Resources Planning and Management
    identifier doi10.1061/JWRMD5.WRENG-6476
    journal fristpage04025012-1
    journal lastpage04025012-10
    page10
    treeJournal of Water Resources Planning and Management:;2025:;Volume ( 151 ):;issue: 006
    contenttypeFulltext
    DSpace software copyright © 2002-2015  DuraSpace
    نرم افزار کتابخانه دیجیتال "دی اسپیس" فارسی شده توسط یابش برای کتابخانه های ایرانی | تماس با یابش
    yabeshDSpacePersian
     
    DSpace software copyright © 2002-2015  DuraSpace
    نرم افزار کتابخانه دیجیتال "دی اسپیس" فارسی شده توسط یابش برای کتابخانه های ایرانی | تماس با یابش
    yabeshDSpacePersian