Attention-Enhanced Multimodal Learning for Conceptual Design Evaluations

被引:9
|
作者
Song, Binyang [1 ]
Miller, Scarlett [2 ]
Ahmed, Faez [1 ]
机构
[1] MIT, Dept Mech Engn, Cambridge, MA 02139 USA
[2] Penn State Univ, Sch Engn Design & Innovat, State Coll, PA 16802 USA
关键词
conceptual design; creativity and concept generation; design evaluation; machine learning; multimodal learning; CREATIVITY; NOVELTY;
D O I
10.1115/1.4056669
中图分类号
TH [机械、仪表工业];
学科分类号
0802 ;
摘要
Conceptual design evaluation is an indispensable component of innovation in the early stage of engineering design. Properly assessing the effectiveness of conceptual design requires a rigorous evaluation of the outputs. Traditional methods to evaluate conceptual designs are slow, expensive, and difficult to scale because they rely on human expert input. An alternative approach is to use computational methods to evaluate design concepts. However, most existing methods have limited utility because they are constrained to unimodal design representations (e.g., texts or sketches). To overcome these limitations, we propose an attention-enhanced multimodal learning (AEMML)-based machine learning (ML) model to predict five design metrics: drawing quality, uniqueness, elegance, usefulness, and creativity. The proposed model utilizes knowledge from large external datasets through transfer learning (TL), simultaneously processes text and sketch data from early-phase concepts, and effectively fuses the multimodal information through a mutual cross-attention mechanism. To study the efficacy of multimodal learning (MML) and attention-based information fusion, we compare (1) a baseline MML model and the unimodal models and (2) the attention-enhanced models with baseline models in terms of their explanatory power for the variability of the design metrics. The results show that MML improves the model explanatory power by 0.05-0.12 and the mutual cross-attention mechanism further increases the explanatory power of the approach by 0.05-0.09, leading to the highest explanatory power of 0.44 for drawing quality, 0.60 for uniqueness, 0.45 for elegance, 0.43 for usefulness, and 0.32 for creativity. Our findings highlight the benefit of using multimodal representations for design metric assessment.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Attention-enhanced and trusted multimodal learning for micro-video venue recognition
    Wang, Bing
    Huang, Xianglin
    Cao, Gang
    Yang, Lifang
    Wei, Xiaolong
    Tao, Zhulin
    [J]. COMPUTERS & ELECTRICAL ENGINEERING, 2022, 102
  • [2] APLNet: Attention-enhanced progressive learning network
    Zhang, Hui
    Kang, Danqing
    He, Haibo
    Wang, Fei-Yue
    [J]. NEUROCOMPUTING, 2020, 371 : 166 - 176
  • [3] Attention-Enhanced Gradual Machine Learning for Entity Resolution
    Zhong, Ping
    Li, Zhanhuai
    Chen, Qun
    Hou, Boyi
    [J]. IEEE INTELLIGENT SYSTEMS, 2021, 36 (06) : 71 - 79
  • [4] Learning Attention-Enhanced Spatiotemporal Representation for Action Recognition
    Shi, Zhensheng
    Cao, Liangjie
    Guan, Cheng
    Zheng, Haiyong
    Gu, Zhaorui
    Yu, Zhibin
    Zheng, Bing
    [J]. IEEE ACCESS, 2020, 8 : 16785 - 16794
  • [5] Attention-enhanced reservoir computing
    Koester, Felix
    Kanno, Kazutaka
    Ohkubo, Jun
    Uchida, Atsushi
    [J]. PHYSICAL REVIEW APPLIED, 2024, 22 (01):
  • [6] ATTENTION-ENHANCED SENSORIMOTOR OBJECT RECOGNITION
    Thermos, Spyridon
    Papadopoulos, Georgios Th.
    Daras, Petros
    Potamianos, Gerasimos
    [J]. 2018 25TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2018, : 336 - 340
  • [7] Personalized Federated Learning Method Based on Attention-Enhanced Meta-Learning Network
    Gao, Yujia
    Wang, Pengfei
    Liu, Liang
    Ma, Huadong
    [J]. Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2024, 61 (01): : 196 - 208
  • [8] Attention-Enhanced Disentangled Representation Learning for Unsupervised Domain Adaptation in Cardiac Segmentation
    Sun, Xiaoyi
    Liu, Zhizhe
    Zheng, Shuai
    Lin, Chen
    Zhu, Zhenfeng
    Zhao, Yao
    [J]. MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2022, PT VII, 2022, 13437 : 745 - 754
  • [9] Attention-enhanced joint learning network for micro-video venue classification
    Bing Wang
    Xianglin Huang
    Gang Cao
    Lifang Yang
    Zhulin Tao
    Xiaolong Wei
    [J]. Multimedia Tools and Applications, 2024, 83 : 12425 - 12443
  • [10] Attention-Enhanced Actor-Critic Learning for Household Nonintrusive Load Monitoring
    Liu, Guohong
    Lv, Liheng
    Wang, Cong
    Wang, Haoming
    Wan, Hui
    Yang, Lijing
    [J]. IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2024,