YaBeSH Engineering and Technology Library

    • Journals
    • PaperQuest
    • YSE Standards
    • YaBeSH
    • Login
    View Item 
    •   YE&T Library
    • ASME
    • Journal of Mechanical Design
    • View Item
    •   YE&T Library
    • ASME
    • Journal of Mechanical Design
    • View Item
    • All Fields
    • Source Title
    • Year
    • Publisher
    • Title
    • Subject
    • Author
    • DOI
    • ISBN
    Advanced Search
    JavaScript is disabled for your browser. Some features of this site may not work without it.

    Archive

    Parametric-ControlNet: Multimodal Control in Foundation Models for Precise Engineering Design Synthesis

    Source: Journal of Mechanical Design:;2025:;volume( 147 ):;issue: 012::page 121704-1
    Author:
    Zhou, Rui
    ,
    Zhang, Yanxia
    ,
    Yuan, Chenyang
    ,
    Permenter, Frank
    ,
    Arechiga, Nikos
    ,
    Klenk, Matt
    ,
    Ahmed, Faez
    DOI: 10.1115/1.4068661
    Publisher: The American Society of Mechanical Engineers (ASME)
    Abstract: This article introduces a generative model designed for multimodal control over text-to-image foundation generative artificial intelligence (AI) models such as Stable Diffusion, specifically tailored for engineering design synthesis. Our model proposes parametric, image, and text control modalities to enhance design precision and diversity. First, it handles both partial and complete parametric inputs using a diffusion model that acts as a design autocomplete copilot, coupled with a parametric encoder to process the information. Second, the model utilizes assembly graphs to systematically assemble input component images, which are then processed through a component encoder to capture essential visual data. Third, textual descriptions are integrated via CLIP encoding, ensuring a comprehensive interpretation of design intent. These diverse inputs are synthesized through a multimodal fusion technique, creating a joint embedding that acts as the input to a module inspired by ControlNet. This integration allows the model to apply robust multimodal control to foundation models, facilitating the generation of complex and precise engineering designs. This approach broadens the capabilities of AI-driven design tools and demonstrates significant advancements in precise control based on diverse data modalities for enhanced design generation.
    • Download: (1.099Mb)
    • Show Full MetaData Hide Full MetaData
    • Get RIS
    • Item Order
    • Go To Publisher
    • Price: 5000 Rial
    • Statistics

      Parametric-ControlNet: Multimodal Control in Foundation Models for Precise Engineering Design Synthesis

    URI
    http://yetl.yabesh.ir/yetl1/handle/yetl/4308026
    Collections
    • Journal of Mechanical Design

    Show full item record

    contributor authorZhou, Rui
    contributor authorZhang, Yanxia
    contributor authorYuan, Chenyang
    contributor authorPermenter, Frank
    contributor authorArechiga, Nikos
    contributor authorKlenk, Matt
    contributor authorAhmed, Faez
    date accessioned2025-08-20T09:17:02Z
    date available2025-08-20T09:17:02Z
    date copyright6/5/2025 12:00:00 AM
    date issued2025
    identifier issn1050-0472
    identifier othermd-24-1872.pdf
    identifier urihttp://yetl.yabesh.ir/yetl1/handle/yetl/4308026
    description abstractThis article introduces a generative model designed for multimodal control over text-to-image foundation generative artificial intelligence (AI) models such as Stable Diffusion, specifically tailored for engineering design synthesis. Our model proposes parametric, image, and text control modalities to enhance design precision and diversity. First, it handles both partial and complete parametric inputs using a diffusion model that acts as a design autocomplete copilot, coupled with a parametric encoder to process the information. Second, the model utilizes assembly graphs to systematically assemble input component images, which are then processed through a component encoder to capture essential visual data. Third, textual descriptions are integrated via CLIP encoding, ensuring a comprehensive interpretation of design intent. These diverse inputs are synthesized through a multimodal fusion technique, creating a joint embedding that acts as the input to a module inspired by ControlNet. This integration allows the model to apply robust multimodal control to foundation models, facilitating the generation of complex and precise engineering designs. This approach broadens the capabilities of AI-driven design tools and demonstrates significant advancements in precise control based on diverse data modalities for enhanced design generation.
    publisherThe American Society of Mechanical Engineers (ASME)
    titleParametric-ControlNet: Multimodal Control in Foundation Models for Precise Engineering Design Synthesis
    typeJournal Paper
    journal volume147
    journal issue12
    journal titleJournal of Mechanical Design
    identifier doi10.1115/1.4068661
    journal fristpage121704-1
    journal lastpage121704-14
    page14
    treeJournal of Mechanical Design:;2025:;volume( 147 ):;issue: 012
    contenttypeFulltext
    DSpace software copyright © 2002-2015  DuraSpace
    نرم افزار کتابخانه دیجیتال "دی اسپیس" فارسی شده توسط یابش برای کتابخانه های ایرانی | تماس با یابش
    yabeshDSpacePersian
     
    DSpace software copyright © 2002-2015  DuraSpace
    نرم افزار کتابخانه دیجیتال "دی اسپیس" فارسی شده توسط یابش برای کتابخانه های ایرانی | تماس با یابش
    yabeshDSpacePersian