YaBeSH Engineering and Technology Library

    • Journals
    • PaperQuest
    • YSE Standards
    • YaBeSH
    • Login
    View Item 
    •   YE&T Library
    • ASME
    • Journal of Computing and Information Science in Engineering
    • View Item
    •   YE&T Library
    • ASME
    • Journal of Computing and Information Science in Engineering
    • View Item
    • All Fields
    • Source Title
    • Year
    • Publisher
    • Title
    • Subject
    • Author
    • DOI
    • ISBN
    Advanced Search
    JavaScript is disabled for your browser. Some features of this site may not work without it.

    Archive

    Epsilon-Greedy Thompson Sampling to Bayesian Optimization

    Source: Journal of Computing and Information Science in Engineering:;2024:;volume( 024 ):;issue: 012::page 121006-1
    Author:
    Do, Bach
    ,
    Adebiyi, Taiwo
    ,
    Zhang, Ruda
    DOI: 10.1115/1.4066858
    Publisher: The American Society of Mechanical Engineers (ASME)
    Abstract: Bayesian optimization (BO) has become a powerful tool for solving simulation-based engineering optimization problems thanks to its ability to integrate physical and mathematical understandings, consider uncertainty, and address the exploitation–exploration dilemma. Thompson sampling (TS) is a preferred solution for BO to handle the exploitation–exploration tradeoff. While it prioritizes exploration by generating and minimizing random sample paths from probabilistic models—a fundamental ingredient of BO–TS weakly manages exploitation by gathering information about the true objective function after it obtains new observations. In this work, we improve the exploitation of TS by incorporating the ε-greedy policy, a well-established selection strategy in reinforcement learning. We first delineate two extremes of TS, namely the generic TS and the sample-average TS. The former promotes exploration, while the latter favors exploitation. We then adopt the ε-greedy policy to randomly switch between these two extremes. Small and large values of ε govern exploitation and exploration, respectively. By minimizing two benchmark functions and solving an inverse problem of a steel cantilever beam, we empirically show that ε-greedy TS equipped with an appropriate ε is more robust than its two extremes, matching or outperforming the better of the generic TS and the sample-average TS.
    • Download: (962.6Kb)
    • Show Full MetaData Hide Full MetaData
    • Get RIS
    • Item Order
    • Go To Publisher
    • Price: 5000 Rial
    • Statistics

      Epsilon-Greedy Thompson Sampling to Bayesian Optimization

    URI
    http://yetl.yabesh.ir/yetl1/handle/yetl/4306129
    Collections
    • Journal of Computing and Information Science in Engineering

    Show full item record

    contributor authorDo, Bach
    contributor authorAdebiyi, Taiwo
    contributor authorZhang, Ruda
    date accessioned2025-04-21T10:24:33Z
    date available2025-04-21T10:24:33Z
    date copyright11/5/2024 12:00:00 AM
    date issued2024
    identifier issn1530-9827
    identifier otherjcise_24_12_121006.pdf
    identifier urihttp://yetl.yabesh.ir/yetl1/handle/yetl/4306129
    description abstractBayesian optimization (BO) has become a powerful tool for solving simulation-based engineering optimization problems thanks to its ability to integrate physical and mathematical understandings, consider uncertainty, and address the exploitation–exploration dilemma. Thompson sampling (TS) is a preferred solution for BO to handle the exploitation–exploration tradeoff. While it prioritizes exploration by generating and minimizing random sample paths from probabilistic models—a fundamental ingredient of BO–TS weakly manages exploitation by gathering information about the true objective function after it obtains new observations. In this work, we improve the exploitation of TS by incorporating the ε-greedy policy, a well-established selection strategy in reinforcement learning. We first delineate two extremes of TS, namely the generic TS and the sample-average TS. The former promotes exploration, while the latter favors exploitation. We then adopt the ε-greedy policy to randomly switch between these two extremes. Small and large values of ε govern exploitation and exploration, respectively. By minimizing two benchmark functions and solving an inverse problem of a steel cantilever beam, we empirically show that ε-greedy TS equipped with an appropriate ε is more robust than its two extremes, matching or outperforming the better of the generic TS and the sample-average TS.
    publisherThe American Society of Mechanical Engineers (ASME)
    titleEpsilon-Greedy Thompson Sampling to Bayesian Optimization
    typeJournal Paper
    journal volume24
    journal issue12
    journal titleJournal of Computing and Information Science in Engineering
    identifier doi10.1115/1.4066858
    journal fristpage121006-1
    journal lastpage121006-10
    page10
    treeJournal of Computing and Information Science in Engineering:;2024:;volume( 024 ):;issue: 012
    contenttypeFulltext
    DSpace software copyright © 2002-2015  DuraSpace
    نرم افزار کتابخانه دیجیتال "دی اسپیس" فارسی شده توسط یابش برای کتابخانه های ایرانی | تماس با یابش
    yabeshDSpacePersian
     
    DSpace software copyright © 2002-2015  DuraSpace
    نرم افزار کتابخانه دیجیتال "دی اسپیس" فارسی شده توسط یابش برای کتابخانه های ایرانی | تماس با یابش
    yabeshDSpacePersian