YaBeSH Engineering and Technology Library

    • Journals
    • PaperQuest
    • YSE Standards
    • YaBeSH
    • Login
    View Item 
    •   YE&T Library
    • ASME
    • ASME Journal of Engineering for Sustainable Buildings and Cities
    • View Item
    •   YE&T Library
    • ASME
    • ASME Journal of Engineering for Sustainable Buildings and Cities
    • View Item
    • All Fields
    • Source Title
    • Year
    • Publisher
    • Title
    • Subject
    • Author
    • DOI
    • ISBN
    Advanced Search
    JavaScript is disabled for your browser. Some features of this site may not work without it.

    Archive

    Optimal Control of District Cooling Energy Plant With Reinforcement Learning and Model Predictive Control

    Source: ASME Journal of Engineering for Sustainable Buildings and Cities:;2023:;volume( 005 ):;issue: 001::page 11002-1
    Author:
    Guo, Zhong
    ,
    Chaudhari, Aditya
    ,
    Coffman, Austin R.
    ,
    Barooah, Prabir
    DOI: 10.1115/1.4064023
    Publisher: The American Society of Mechanical Engineers (ASME)
    Abstract: We consider the problem of optimal control of district cooling energy plants (DCEPs) consisting of multiple chillers, a cooling tower, and a thermal energy storage (TES), in the presence of time-varying electricity prices. A straightforward application of model predictive control (MPC) requires solving a challenging mixed-integer nonlinear program (MINLP) because of the on/off of chillers and the complexity of the DCEP model. Reinforcement learning (RL) is an attractive alternative since its real-time control computation is much simpler. But designing an RL controller is challenging due to myriad design choices and computationally intensive training. In this paper, we propose an RL controller and an MPC controller for minimizing the electricity cost of a DCEP, and compare them via simulations. The two controllers are designed to be comparable in terms of objective and information requirements. The RL controller uses a novel Q-learning algorithm that is based on least-squares policy iteration. We describe the design choices for the RL controller, including the choice of state space and basis functions, that are found to be effective. The proposed MPC controller does not need a mixed-integer solver for implementation, but only a nonlinear program (NLP) solver. A rule-based baseline controller is also proposed to aid in comparison. Simulation results show that the proposed RL and MPC controllers achieve similar savings over the baseline controller, about 17%.
    • Download: (991.5Kb)
    • Show Full MetaData Hide Full MetaData
    • Get RIS
    • Item Order
    • Go To Publisher
    • Price: 5000 Rial
    • Statistics

      Optimal Control of District Cooling Energy Plant With Reinforcement Learning and Model Predictive Control

    URI
    http://yetl.yabesh.ir/yetl1/handle/yetl/4295503
    Collections
    • ASME Journal of Engineering for Sustainable Buildings and Cities

    Show full item record

    contributor authorGuo, Zhong
    contributor authorChaudhari, Aditya
    contributor authorCoffman, Austin R.
    contributor authorBarooah, Prabir
    date accessioned2024-04-24T22:35:44Z
    date available2024-04-24T22:35:44Z
    date copyright12/4/2023 12:00:00 AM
    date issued2023
    identifier issn2642-6641
    identifier otherjesbc_5_1_011002.pdf
    identifier urihttp://yetl.yabesh.ir/yetl1/handle/yetl/4295503
    description abstractWe consider the problem of optimal control of district cooling energy plants (DCEPs) consisting of multiple chillers, a cooling tower, and a thermal energy storage (TES), in the presence of time-varying electricity prices. A straightforward application of model predictive control (MPC) requires solving a challenging mixed-integer nonlinear program (MINLP) because of the on/off of chillers and the complexity of the DCEP model. Reinforcement learning (RL) is an attractive alternative since its real-time control computation is much simpler. But designing an RL controller is challenging due to myriad design choices and computationally intensive training. In this paper, we propose an RL controller and an MPC controller for minimizing the electricity cost of a DCEP, and compare them via simulations. The two controllers are designed to be comparable in terms of objective and information requirements. The RL controller uses a novel Q-learning algorithm that is based on least-squares policy iteration. We describe the design choices for the RL controller, including the choice of state space and basis functions, that are found to be effective. The proposed MPC controller does not need a mixed-integer solver for implementation, but only a nonlinear program (NLP) solver. A rule-based baseline controller is also proposed to aid in comparison. Simulation results show that the proposed RL and MPC controllers achieve similar savings over the baseline controller, about 17%.
    publisherThe American Society of Mechanical Engineers (ASME)
    titleOptimal Control of District Cooling Energy Plant With Reinforcement Learning and Model Predictive Control
    typeJournal Paper
    journal volume5
    journal issue1
    journal titleASME Journal of Engineering for Sustainable Buildings and Cities
    identifier doi10.1115/1.4064023
    journal fristpage11002-1
    journal lastpage11002-12
    page12
    treeASME Journal of Engineering for Sustainable Buildings and Cities:;2023:;volume( 005 ):;issue: 001
    contenttypeFulltext
    DSpace software copyright © 2002-2015  DuraSpace
    نرم افزار کتابخانه دیجیتال "دی اسپیس" فارسی شده توسط یابش برای کتابخانه های ایرانی | تماس با یابش
    yabeshDSpacePersian
     
    DSpace software copyright © 2002-2015  DuraSpace
    نرم افزار کتابخانه دیجیتال "دی اسپیس" فارسی شده توسط یابش برای کتابخانه های ایرانی | تماس با یابش
    yabeshDSpacePersian