YaBeSH Engineering and Technology Library

    • Journals
    • PaperQuest
    • YSE Standards
    • YaBeSH
    • Login
    View Item 
    •   YE&T Library
    • ASME
    • Journal of Manufacturing Science and Engineering
    • View Item
    •   YE&T Library
    • ASME
    • Journal of Manufacturing Science and Engineering
    • View Item
    • All Fields
    • Source Title
    • Year
    • Publisher
    • Title
    • Subject
    • Author
    • DOI
    • ISBN
    Advanced Search
    JavaScript is disabled for your browser. Some features of this site may not work without it.

    Archive

    Integrated Energy Optimization in Manufacturing Through Multiagent Deep Reinforcement Learning: Holistic Control of Manufacturing, Microgrid Systems, and Battery Storage

    Source: Journal of Manufacturing Science and Engineering:;2025:;volume( 147 ):;issue: 006::page 61003-1
    Author:
    Waseem, Muhammad
    ,
    Maithripala, Mihitha Sarinda
    ,
    Chang, Qing
    ,
    Lin, Zongli
    DOI: 10.1115/1.4067614
    Publisher: The American Society of Mechanical Engineers (ASME)
    Abstract: Microgrid technology integrates storage devices, renewable energy sources, and controllable loads and has been widely explored in residential, commercial, and critical facilities. However, its potential in manufacturing remains largely underexplored, where optimal control of microgrids containing energy storage systems (ESS) is crucial. Two primary challenges arise in integrated microgrid-manufacturing systems: fluctuating renewable energy output and nondeterministic polynomial (NP)-hard demand-side control. Addressing both challenges simultaneously increases complexity. This article proposes an integrated control considering ESS degradation, optimizing control on both the manufacturing demand and microgrid energy supply sides within the production constraints. It formulates the problem in a decentralized partially observable Markov decision process (Dec-POMDP) framework, treating the system as a multiagent environment. The multiagent deep deterministic policy gradient (MADDPG) algorithm is adapted to optimize control policies. Investigating the trained policies provides insights into their logic, and a rule-based policy is introduced for practical implementation. Experimental validation on a manufacturing system validates the effectiveness of the proposed method and the rule-based policy.
    • Download: (1.184Mb)
    • Show Full MetaData Hide Full MetaData
    • Get RIS
    • Item Order
    • Go To Publisher
    • Price: 5000 Rial
    • Statistics

      Integrated Energy Optimization in Manufacturing Through Multiagent Deep Reinforcement Learning: Holistic Control of Manufacturing, Microgrid Systems, and Battery Storage

    URI
    http://yetl.yabesh.ir/yetl1/handle/yetl/4305394
    Collections
    • Journal of Manufacturing Science and Engineering

    Show full item record

    contributor authorWaseem, Muhammad
    contributor authorMaithripala, Mihitha Sarinda
    contributor authorChang, Qing
    contributor authorLin, Zongli
    date accessioned2025-04-21T10:03:17Z
    date available2025-04-21T10:03:17Z
    date copyright2/11/2025 12:00:00 AM
    date issued2025
    identifier issn1087-1357
    identifier othermanu-24-1440.pdf
    identifier urihttp://yetl.yabesh.ir/yetl1/handle/yetl/4305394
    description abstractMicrogrid technology integrates storage devices, renewable energy sources, and controllable loads and has been widely explored in residential, commercial, and critical facilities. However, its potential in manufacturing remains largely underexplored, where optimal control of microgrids containing energy storage systems (ESS) is crucial. Two primary challenges arise in integrated microgrid-manufacturing systems: fluctuating renewable energy output and nondeterministic polynomial (NP)-hard demand-side control. Addressing both challenges simultaneously increases complexity. This article proposes an integrated control considering ESS degradation, optimizing control on both the manufacturing demand and microgrid energy supply sides within the production constraints. It formulates the problem in a decentralized partially observable Markov decision process (Dec-POMDP) framework, treating the system as a multiagent environment. The multiagent deep deterministic policy gradient (MADDPG) algorithm is adapted to optimize control policies. Investigating the trained policies provides insights into their logic, and a rule-based policy is introduced for practical implementation. Experimental validation on a manufacturing system validates the effectiveness of the proposed method and the rule-based policy.
    publisherThe American Society of Mechanical Engineers (ASME)
    titleIntegrated Energy Optimization in Manufacturing Through Multiagent Deep Reinforcement Learning: Holistic Control of Manufacturing, Microgrid Systems, and Battery Storage
    typeJournal Paper
    journal volume147
    journal issue6
    journal titleJournal of Manufacturing Science and Engineering
    identifier doi10.1115/1.4067614
    journal fristpage61003-1
    journal lastpage61003-14
    page14
    treeJournal of Manufacturing Science and Engineering:;2025:;volume( 147 ):;issue: 006
    contenttypeFulltext
    DSpace software copyright © 2002-2015  DuraSpace
    نرم افزار کتابخانه دیجیتال "دی اسپیس" فارسی شده توسط یابش برای کتابخانه های ایرانی | تماس با یابش
    yabeshDSpacePersian
     
    DSpace software copyright © 2002-2015  DuraSpace
    نرم افزار کتابخانه دیجیتال "دی اسپیس" فارسی شده توسط یابش برای کتابخانه های ایرانی | تماس با یابش
    yabeshDSpacePersian