TY - GEN
T1 - A PICTURE OR A THOUSAND WORDS
T2 - ASME 2024 International Design Engineering Technical Conferences and Computers and Information in Engineering Conference, IDETC-CIE 2024
AU - Keeler, Matthew
AU - Fuge, Mark
AU - Peng, Aoran
AU - Miller, Scarlett
N1 - Publisher Copyright:
Copyright © 2024 by ASME.
PY - 2024
Y1 - 2024
N2 - Well-studied techniques that enhance diversity in early design concept generation require effective metrics for evaluating human-perceived similarity between ideas. Recent work suggests collecting triplet comparisons between designs directly from human raters and using those triplets to form an embedding where similarity is expressed as a Euclidean distance. While effective at modeling human-perceived similarity judgments, these methods are expensive and require a large number of triplets to be hand-labeled. However, what if there were a way to use AI to replicate the human similarity judgments captured in triplet embedding methods? In this paper, we explore the potential for pretrained Large Language Models (LLMs) to be used in this context. Using a dataset of crowdsourced text descriptions written about engineering design sketches, we generate LLM embeddings and compare them to an embedding created from human-provided triplets of those same sketches. From these embeddings, we can use Euclidean distances to describe areas where human perception and LLM perception disagree regarding design similarity. We then implement this same procedure but with descriptions written from a template that attempts to isolate a particular modality of a design (i.e., functions, behaviors, structures). By comparing the templated description embeddings to both the triplet-generated and pre-template LLM embeddings, we identify ways of describing designs such that LLM and human similarity perception better agree. We use these results to better understand how humans and LLMs interpret similarity in engineering designs.
AB - Well-studied techniques that enhance diversity in early design concept generation require effective metrics for evaluating human-perceived similarity between ideas. Recent work suggests collecting triplet comparisons between designs directly from human raters and using those triplets to form an embedding where similarity is expressed as a Euclidean distance. While effective at modeling human-perceived similarity judgments, these methods are expensive and require a large number of triplets to be hand-labeled. However, what if there were a way to use AI to replicate the human similarity judgments captured in triplet embedding methods? In this paper, we explore the potential for pretrained Large Language Models (LLMs) to be used in this context. Using a dataset of crowdsourced text descriptions written about engineering design sketches, we generate LLM embeddings and compare them to an embedding created from human-provided triplets of those same sketches. From these embeddings, we can use Euclidean distances to describe areas where human perception and LLM perception disagree regarding design similarity. We then implement this same procedure but with descriptions written from a template that attempts to isolate a particular modality of a design (i.e., functions, behaviors, structures). By comparing the templated description embeddings to both the triplet-generated and pre-template LLM embeddings, we identify ways of describing designs such that LLM and human similarity perception better agree. We use these results to better understand how humans and LLMs interpret similarity in engineering designs.
UR - http://www.scopus.com/inward/record.url?scp=85210258937&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85210258937&partnerID=8YFLogxK
U2 - 10.1115/DETC2024-143634
DO - 10.1115/DETC2024-143634
M3 - Conference contribution
AN - SCOPUS:85210258937
T3 - Proceedings of the ASME Design Engineering Technical Conference
BT - 36th International Conference on Design Theory and Methodology (DTM)
PB - American Society of Mechanical Engineers (ASME)
Y2 - 25 August 2024 through 28 August 2024
ER -