Show simple item record

contributor authorTao, Siyu
contributor authorSharma, Chaitra
contributor authorDevanathan, Srikanth
date accessioned2024-12-24T18:47:01Z
date available2024-12-24T18:47:01Z
date copyright8/2/2024 12:00:00 AM
date issued2024
identifier issn2377-2158
identifier othervvuq_009_02_021010.pdf
identifier urihttp://yetl.yabesh.ir/yetl1/handle/yetl/4302735
description abstractMultifidelity (MF) models abound in simulation-based engineering. Many MF strategies have been proposed to improve the efficiency in engineering processes, especially in design optimization. When it comes to assessing the performance of MF optimization techniques, existing practice usually relies on test cases involving contrived MF models of seemingly random math functions, due to limited access to real-world MF models. While it is acceptable to use contrived MF models, these models are often manually written up rather than created in a systematic manner. This gives rise to the potential pitfall that the test MF models may be not representative of general scenarios. We propose a framework to generate test MF models systematically and characterize MF optimization techniques' performances comprehensively. In our framework, the MF models are generated based on given high-fidelity (HF) models and come with two parameters to control their fidelity levels and allow model randomization. In our testing process, MF case problems are systematically formulated using our model creation method. Running the given MF optimization technique on these problems produces what we call “savings curves” that characterize the technique's performance similarly to how receiver operating characteristic (ROC) curves characterize machine learning classifiers. Our test results also allow plotting “optimality curves” that serve similar functionality to savings curves in certain types of problems. The flexibility of our MF model creation facilitates the development of testing processes for general MF problem scenarios, and our framework can be easily extended to other MF applications than optimization.
publisherThe American Society of Mechanical Engineers (ASME)
titleA Framework for Developing Systematic Testbeds for Multifidelity Optimization Techniques
typeJournal Paper
journal volume9
journal issue2
journal titleJournal of Verification, Validation and Uncertainty Quantification
identifier doi10.1115/1.4065719
journal fristpage21010-1
journal lastpage21010-9
page9
treeJournal of Verification, Validation and Uncertainty Quantification:;2024:;volume( 009 ):;issue: 002
contenttypeFulltext


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record