YaBeSH Engineering and Technology Library

    • Journals
    • PaperQuest
    • YSE Standards
    • YaBeSH
    • Login
    View Item 
    •   YE&T Library
    • ASME
    • Journal of Dynamic Systems, Measurement, and Control
    • View Item
    •   YE&T Library
    • ASME
    • Journal of Dynamic Systems, Measurement, and Control
    • View Item
    • All Fields
    • Source Title
    • Year
    • Publisher
    • Title
    • Subject
    • Author
    • DOI
    • ISBN
    Advanced Search
    JavaScript is disabled for your browser. Some features of this site may not work without it.

    Archive

    Inferring Human Control Intent Using Inverse Linear Quadratic Regulator With Output Penalty Versus Gain Penalty: Better Fit but Similar Intent

    Source: Journal of Dynamic Systems, Measurement, and Control:;2024:;volume( 146 ):;issue: 006::page 61103-1
    Author:
    Yu, Heejin
    ,
    Ramadan, Ahmed
    ,
    Cholewicki, Jacek
    ,
    Popovich, John M., Jr.
    ,
    Reeves, N. Peter
    ,
    You, Joshua (Sung) H.
    ,
    Choi, Jongeun
    DOI: 10.1115/1.4065593
    Publisher: The American Society of Mechanical Engineers (ASME)
    Abstract: To identify the underlying mechanisms of human motor control, parametric models are utilized. One approach of employing these models is the inferring the control intent (estimating motor control strategy). A well-accepted assumption is that human motor control is optimal; thus, the intent is inferred by solving an inverse optimal control (IOC) problem. Linear quadratic regulator (LQR) is a well-established optimal controller, and its inverse LQR (ILQR) problem has been used in the literature to infer the control intent of one subject. This implementation used a cost function with gain penalty, minimizing the error between LQR gain and a preliminary estimated gain. We hypothesize that relying on an estimated gain may limit ILQR optimization capability. In this study, we derive an ILQR optimization with output penalty, minimizing the error between the model output and the measured output. We conducted the test on 30 healthy subjects who sat on a robotic seat capable of rotation. The task utilized a physical human–robot interaction with a perturbation torque as input and lower and upper body angles as output. Our method significantly improved the goodness of fit compared to the gain-penalty ILQR. Moreover, the dominant inferred intent was not statistically different between the two methods. To our knowledge, this work is the first that infers motor control intent for a sample of healthy subjects. This is a step closer to investigating control intent differences between healthy subjects and subjects with altered motor control, e.g., low back pain.
    • Download: (1.083Mb)
    • Show Full MetaData Hide Full MetaData
    • Get RIS
    • Item Order
    • Go To Publisher
    • Price: 5000 Rial
    • Statistics

      Inferring Human Control Intent Using Inverse Linear Quadratic Regulator With Output Penalty Versus Gain Penalty: Better Fit but Similar Intent

    URI
    http://yetl.yabesh.ir/yetl1/handle/yetl/4302815
    Collections
    • Journal of Dynamic Systems, Measurement, and Control

    Show full item record

    contributor authorYu, Heejin
    contributor authorRamadan, Ahmed
    contributor authorCholewicki, Jacek
    contributor authorPopovich, John M., Jr.
    contributor authorReeves, N. Peter
    contributor authorYou, Joshua (Sung) H.
    contributor authorChoi, Jongeun
    date accessioned2024-12-24T18:49:25Z
    date available2024-12-24T18:49:25Z
    date copyright7/16/2024 12:00:00 AM
    date issued2024
    identifier issn0022-0434
    identifier otherds_146_06_061103.pdf
    identifier urihttp://yetl.yabesh.ir/yetl1/handle/yetl/4302815
    description abstractTo identify the underlying mechanisms of human motor control, parametric models are utilized. One approach of employing these models is the inferring the control intent (estimating motor control strategy). A well-accepted assumption is that human motor control is optimal; thus, the intent is inferred by solving an inverse optimal control (IOC) problem. Linear quadratic regulator (LQR) is a well-established optimal controller, and its inverse LQR (ILQR) problem has been used in the literature to infer the control intent of one subject. This implementation used a cost function with gain penalty, minimizing the error between LQR gain and a preliminary estimated gain. We hypothesize that relying on an estimated gain may limit ILQR optimization capability. In this study, we derive an ILQR optimization with output penalty, minimizing the error between the model output and the measured output. We conducted the test on 30 healthy subjects who sat on a robotic seat capable of rotation. The task utilized a physical human–robot interaction with a perturbation torque as input and lower and upper body angles as output. Our method significantly improved the goodness of fit compared to the gain-penalty ILQR. Moreover, the dominant inferred intent was not statistically different between the two methods. To our knowledge, this work is the first that infers motor control intent for a sample of healthy subjects. This is a step closer to investigating control intent differences between healthy subjects and subjects with altered motor control, e.g., low back pain.
    publisherThe American Society of Mechanical Engineers (ASME)
    titleInferring Human Control Intent Using Inverse Linear Quadratic Regulator With Output Penalty Versus Gain Penalty: Better Fit but Similar Intent
    typeJournal Paper
    journal volume146
    journal issue6
    journal titleJournal of Dynamic Systems, Measurement, and Control
    identifier doi10.1115/1.4065593
    journal fristpage61103-1
    journal lastpage61103-7
    page7
    treeJournal of Dynamic Systems, Measurement, and Control:;2024:;volume( 146 ):;issue: 006
    contenttypeFulltext
    DSpace software copyright © 2002-2015  DuraSpace
    نرم افزار کتابخانه دیجیتال "دی اسپیس" فارسی شده توسط یابش برای کتابخانه های ایرانی | تماس با یابش
    yabeshDSpacePersian
     
    DSpace software copyright © 2002-2015  DuraSpace
    نرم افزار کتابخانه دیجیتال "دی اسپیس" فارسی شده توسط یابش برای کتابخانه های ایرانی | تماس با یابش
    yabeshDSpacePersian