YaBeSH Engineering and Technology Library

    • Journals
    • PaperQuest
    • YSE Standards
    • YaBeSH
    • Login
    View Item 
    •   YE&T Library
    • ASCE
    • Journal of Computing in Civil Engineering
    • View Item
    •   YE&T Library
    • ASCE
    • Journal of Computing in Civil Engineering
    • View Item
    • All Fields
    • Source Title
    • Year
    • Publisher
    • Title
    • Subject
    • Author
    • DOI
    • ISBN
    Advanced Search
    JavaScript is disabled for your browser. Some features of this site may not work without it.

    Archive

    AdaLN: A Vision Transformer for Multidomain Learning and Predisaster Building Information Extraction from Images

    Source: Journal of Computing in Civil Engineering:;2022:;Volume ( 036 ):;issue: 005::page 04022024
    Author:
    Yunhui Guo
    ,
    Chaofeng Wang
    ,
    Stella X. Yu
    ,
    Frank McKenna
    ,
    Kincho H. Law
    DOI: 10.1061/(ASCE)CP.1943-5487.0001034
    Publisher: ASCE
    Abstract: Satellite and street view images are widely used in various disciplines as a source of information for understanding the built environment. In natural hazard engineering, high-quality building inventory data sets are crucial for the simulation of hazard impacts and for supporting decision-making. Screening the building stocks to gather the information for simulation and to detect potential structural defects that are vulnerable to natural hazards is a time-consuming and labor-intensive task. This paper presents an automated method for extracting building information through the use of satellite and street view images. The method is built upon a novel transformer-based deep neural network we developed. Specifically, a multidomain learning approach is employed to develop a single compact model for multiple image-based deep learning information extraction tasks using multiple data sources (e.g., satellite and street view images). Our multidomain Vision Transformer is designed as a unified architecture that can be effectively deployed for multiple classification tasks. The effectiveness of the proposed approach is demonstrated in a case study in which we use pretrained models to collect regional-scale building information that is related to natural hazard risks.
    • Download: (2.666Mb)
    • Show Full MetaData Hide Full MetaData
    • Get RIS
    • Item Order
    • Go To Publisher
    • Price: 5000 Rial
    • Statistics

      AdaLN: A Vision Transformer for Multidomain Learning and Predisaster Building Information Extraction from Images

    URI
    http://yetl.yabesh.ir/yetl1/handle/yetl/4286178
    Collections
    • Journal of Computing in Civil Engineering

    Show full item record

    contributor authorYunhui Guo
    contributor authorChaofeng Wang
    contributor authorStella X. Yu
    contributor authorFrank McKenna
    contributor authorKincho H. Law
    date accessioned2022-08-18T12:11:44Z
    date available2022-08-18T12:11:44Z
    date issued2022/07/04
    identifier other%28ASCE%29CP.1943-5487.0001034.pdf
    identifier urihttp://yetl.yabesh.ir/yetl1/handle/yetl/4286178
    description abstractSatellite and street view images are widely used in various disciplines as a source of information for understanding the built environment. In natural hazard engineering, high-quality building inventory data sets are crucial for the simulation of hazard impacts and for supporting decision-making. Screening the building stocks to gather the information for simulation and to detect potential structural defects that are vulnerable to natural hazards is a time-consuming and labor-intensive task. This paper presents an automated method for extracting building information through the use of satellite and street view images. The method is built upon a novel transformer-based deep neural network we developed. Specifically, a multidomain learning approach is employed to develop a single compact model for multiple image-based deep learning information extraction tasks using multiple data sources (e.g., satellite and street view images). Our multidomain Vision Transformer is designed as a unified architecture that can be effectively deployed for multiple classification tasks. The effectiveness of the proposed approach is demonstrated in a case study in which we use pretrained models to collect regional-scale building information that is related to natural hazard risks.
    publisherASCE
    titleAdaLN: A Vision Transformer for Multidomain Learning and Predisaster Building Information Extraction from Images
    typeJournal Article
    journal volume36
    journal issue5
    journal titleJournal of Computing in Civil Engineering
    identifier doi10.1061/(ASCE)CP.1943-5487.0001034
    journal fristpage04022024
    journal lastpage04022024-12
    page12
    treeJournal of Computing in Civil Engineering:;2022:;Volume ( 036 ):;issue: 005
    contenttypeFulltext
    DSpace software copyright © 2002-2015  DuraSpace
    نرم افزار کتابخانه دیجیتال "دی اسپیس" فارسی شده توسط یابش برای کتابخانه های ایرانی | تماس با یابش
    yabeshDSpacePersian
     
    DSpace software copyright © 2002-2015  DuraSpace
    نرم افزار کتابخانه دیجیتال "دی اسپیس" فارسی شده توسط یابش برای کتابخانه های ایرانی | تماس با یابش
    yabeshDSpacePersian