YaBeSH Engineering and Technology Library

    • Journals
    • PaperQuest
    • YSE Standards
    • YaBeSH
    • Login
    View Item 
    •   YE&T Library
    • ASME
    • Journal of Dynamic Systems, Measurement, and Control
    • View Item
    •   YE&T Library
    • ASME
    • Journal of Dynamic Systems, Measurement, and Control
    • View Item
    • All Fields
    • Source Title
    • Year
    • Publisher
    • Title
    • Subject
    • Author
    • DOI
    • ISBN
    Advanced Search
    JavaScript is disabled for your browser. Some features of this site may not work without it.

    Archive

    A Real-Time Computational Learning Model for Sequential Decision-Making Problems Under Uncertainty

    Source: Journal of Dynamic Systems, Measurement, and Control:;2009:;volume( 131 ):;issue: 004::page 41010
    Author:
    Andreas A. Malikopoulos
    ,
    Panos Y. Papalambros
    ,
    Dennis N. Assanis
    DOI: 10.1115/1.3117200
    Publisher: The American Society of Mechanical Engineers (ASME)
    Abstract: Modeling dynamic systems incurring stochastic disturbances for deriving a control policy is a ubiquitous task in engineering. However, in some instances obtaining a model of a system may be impractical or impossible. Alternative approaches have been developed using a simulation-based stochastic framework, in which the system interacts with its environment in real time and obtains information that can be processed to produce an optimal control policy. In this context, the problem of developing a policy for controlling the system’s behavior is formulated as a sequential decision-making problem under uncertainty. This paper considers the problem of deriving a control policy for a dynamic system with unknown dynamics in real time, formulated as a sequential decision-making under uncertainty. The evolution of the system is modeled as a controlled Markov chain. A new state-space representation model and a learning mechanism are proposed that can be used to improve system performance over time. The major difference between the existing methods and the proposed learning model is that the latter utilizes an evaluation function, which considers the expected cost that can be achieved by state transitions forward in time. The model allows decision-making based on gradually enhanced knowledge of system response as it transitions from one state to another, in conjunction with actions taken at each state. The proposed model is demonstrated on the single cart-pole balancing problem and a vehicle cruise-control problem.
    • Download: (687.4Kb)
    • Show Full MetaData Hide Full MetaData
    • Get RIS
    • Item Order
    • Go To Publisher
    • Price: 5000 Rial
    • Statistics

      A Real-Time Computational Learning Model for Sequential Decision-Making Problems Under Uncertainty

    URI
    http://yetl.yabesh.ir/yetl1/handle/yetl/140199
    Collections
    • Journal of Dynamic Systems, Measurement, and Control

    Show full item record

    contributor authorAndreas A. Malikopoulos
    contributor authorPanos Y. Papalambros
    contributor authorDennis N. Assanis
    date accessioned2017-05-09T00:32:10Z
    date available2017-05-09T00:32:10Z
    date copyrightJuly, 2009
    date issued2009
    identifier issn0022-0434
    identifier otherJDSMAA-26497#041010_1.pdf
    identifier urihttp://yetl.yabesh.ir/yetl/handle/yetl/140199
    description abstractModeling dynamic systems incurring stochastic disturbances for deriving a control policy is a ubiquitous task in engineering. However, in some instances obtaining a model of a system may be impractical or impossible. Alternative approaches have been developed using a simulation-based stochastic framework, in which the system interacts with its environment in real time and obtains information that can be processed to produce an optimal control policy. In this context, the problem of developing a policy for controlling the system’s behavior is formulated as a sequential decision-making problem under uncertainty. This paper considers the problem of deriving a control policy for a dynamic system with unknown dynamics in real time, formulated as a sequential decision-making under uncertainty. The evolution of the system is modeled as a controlled Markov chain. A new state-space representation model and a learning mechanism are proposed that can be used to improve system performance over time. The major difference between the existing methods and the proposed learning model is that the latter utilizes an evaluation function, which considers the expected cost that can be achieved by state transitions forward in time. The model allows decision-making based on gradually enhanced knowledge of system response as it transitions from one state to another, in conjunction with actions taken at each state. The proposed model is demonstrated on the single cart-pole balancing problem and a vehicle cruise-control problem.
    publisherThe American Society of Mechanical Engineers (ASME)
    titleA Real-Time Computational Learning Model for Sequential Decision-Making Problems Under Uncertainty
    typeJournal Paper
    journal volume131
    journal issue4
    journal titleJournal of Dynamic Systems, Measurement, and Control
    identifier doi10.1115/1.3117200
    journal fristpage41010
    identifier eissn1528-9028
    treeJournal of Dynamic Systems, Measurement, and Control:;2009:;volume( 131 ):;issue: 004
    contenttypeFulltext
    DSpace software copyright © 2002-2015  DuraSpace
    نرم افزار کتابخانه دیجیتال "دی اسپیس" فارسی شده توسط یابش برای کتابخانه های ایرانی | تماس با یابش
    yabeshDSpacePersian
     
    DSpace software copyright © 2002-2015  DuraSpace
    نرم افزار کتابخانه دیجیتال "دی اسپیس" فارسی شده توسط یابش برای کتابخانه های ایرانی | تماس با یابش
    yabeshDSpacePersian