YaBeSH Engineering and Technology Library

    • Journals
    • PaperQuest
    • YSE Standards
    • YaBeSH
    • Login
    View Item 
    •   YE&T Library
    • ASME
    • Journal of Manufacturing Science and Engineering
    • View Item
    •   YE&T Library
    • ASME
    • Journal of Manufacturing Science and Engineering
    • View Item
    • All Fields
    • Source Title
    • Year
    • Publisher
    • Title
    • Subject
    • Author
    • DOI
    • ISBN
    Advanced Search
    JavaScript is disabled for your browser. Some features of this site may not work without it.

    Archive

    Deep Reinforcement Learning-Based Multi-Task Scheduling in Cloud Manufacturing Under Different Task Arrival Modes

    Source: Journal of Manufacturing Science and Engineering:;2023:;volume( 145 ):;issue: 008::page 81003-1
    Author:
    Ping, Yaoyao
    ,
    Liu, Yongkui
    ,
    Zhang, Lin
    ,
    Wang, Lihui
    ,
    Xu, Xun
    DOI: 10.1115/1.4062217
    Publisher: The American Society of Mechanical Engineers (ASME)
    Abstract: Cloud manufacturing is a service-oriented networked manufacturing model that aims to provide manufacturing resources as services in an on-demand manner. Scheduling is one of the key techniques for cloud manufacturing to achieve the aim. Multi-task scheduling with dynamical task arrivals is a critical problem in cloud manufacturing. Many traditional algorithms such as the genetic algorithm (GA) and ant colony optimization algorithm (ACO) have been used to address the issue, which, however, either are incapable of or perform poorly in tackling the problem. Deep reinforcement learning (DRL) as the combination of deep learning (DL) and reinforcement learning (RL) provides an effective technique in this regard. In view of this, we employ a typical DRL algorithm—Deep Q-network (DQN)—and propose a DQN-based approach for multitask scheduling in cloud manufacturing. Three different task arrival modes—arriving at the same time, arriving in random batches, and arriving one by one sequentially—are considered. Four baseline methods including random scheduling, round-robin scheduling, earliest scheduling, and minimum execution time (min-time) scheduling are investigated. A comparison of results indicates that the DQN-based scheduling approach is effective and performs best among all approaches in addressing the multitask scheduling problem in cloud manufacturing.
    • Download: (1.230Mb)
    • Show Full MetaData Hide Full MetaData
    • Get RIS
    • Item Order
    • Go To Publisher
    • Price: 5000 Rial
    • Statistics

      Deep Reinforcement Learning-Based Multi-Task Scheduling in Cloud Manufacturing Under Different Task Arrival Modes

    URI
    http://yetl.yabesh.ir/yetl1/handle/yetl/4294754
    Collections
    • Journal of Manufacturing Science and Engineering

    Show full item record

    contributor authorPing, Yaoyao
    contributor authorLiu, Yongkui
    contributor authorZhang, Lin
    contributor authorWang, Lihui
    contributor authorXu, Xun
    date accessioned2023-11-29T19:25:55Z
    date available2023-11-29T19:25:55Z
    date copyright4/12/2023 12:00:00 AM
    date issued4/12/2023 12:00:00 AM
    date issued2023-04-12
    identifier issn1087-1357
    identifier othermanu_145_8_081003.pdf
    identifier urihttp://yetl.yabesh.ir/yetl1/handle/yetl/4294754
    description abstractCloud manufacturing is a service-oriented networked manufacturing model that aims to provide manufacturing resources as services in an on-demand manner. Scheduling is one of the key techniques for cloud manufacturing to achieve the aim. Multi-task scheduling with dynamical task arrivals is a critical problem in cloud manufacturing. Many traditional algorithms such as the genetic algorithm (GA) and ant colony optimization algorithm (ACO) have been used to address the issue, which, however, either are incapable of or perform poorly in tackling the problem. Deep reinforcement learning (DRL) as the combination of deep learning (DL) and reinforcement learning (RL) provides an effective technique in this regard. In view of this, we employ a typical DRL algorithm—Deep Q-network (DQN)—and propose a DQN-based approach for multitask scheduling in cloud manufacturing. Three different task arrival modes—arriving at the same time, arriving in random batches, and arriving one by one sequentially—are considered. Four baseline methods including random scheduling, round-robin scheduling, earliest scheduling, and minimum execution time (min-time) scheduling are investigated. A comparison of results indicates that the DQN-based scheduling approach is effective and performs best among all approaches in addressing the multitask scheduling problem in cloud manufacturing.
    publisherThe American Society of Mechanical Engineers (ASME)
    titleDeep Reinforcement Learning-Based Multi-Task Scheduling in Cloud Manufacturing Under Different Task Arrival Modes
    typeJournal Paper
    journal volume145
    journal issue8
    journal titleJournal of Manufacturing Science and Engineering
    identifier doi10.1115/1.4062217
    journal fristpage81003-1
    journal lastpage81003-12
    page12
    treeJournal of Manufacturing Science and Engineering:;2023:;volume( 145 ):;issue: 008
    contenttypeFulltext
    DSpace software copyright © 2002-2015  DuraSpace
    نرم افزار کتابخانه دیجیتال "دی اسپیس" فارسی شده توسط یابش برای کتابخانه های ایرانی | تماس با یابش
    yabeshDSpacePersian
     
    DSpace software copyright © 2002-2015  DuraSpace
    نرم افزار کتابخانه دیجیتال "دی اسپیس" فارسی شده توسط یابش برای کتابخانه های ایرانی | تماس با یابش
    yabeshDSpacePersian