Show simple item record

contributor authorZhang, Xinyao
contributor authorTian, Sibo
contributor authorLiang, Xiao
contributor authorZheng, Minghui
contributor authorBehdad, Sara
date accessioned2025-08-20T09:30:28Z
date available2025-08-20T09:30:28Z
date copyright3/10/2025 12:00:00 AM
date issued2025
identifier issn1530-9827
identifier otherjcise-24-1293.pdf
identifier urihttp://yetl.yabesh.ir/yetl1/handle/yetl/4308389
description abstractHuman–robot collaboration (HRC) has become an integral element of many manufacturing and service industries. A fundamental requirement for safe HRC is understanding and predicting human trajectories and intentions, especially when humans and robots operate nearby. Although existing research emphasizes predicting human motions or intentions, a key challenge is predicting both human trajectories and intentions simultaneously. This paper addresses this gap by developing a multi-task learning framework consisting of a bi-long short-term memory-based encoder–decoder architecture that obtains the motion data from both human and robot trajectories as inputs and performs two main tasks simultaneously: human trajectory prediction and human intention prediction. The first task predicts human trajectories by reconstructing the motion sequences, while the second task tests two main approaches for intention prediction: supervised learning, specifically a support vector machine, to predict human intention based on the latent representation, and, an unsupervised learning method, the hidden Markov model, that decodes the latent features for human intention prediction. Four encoder designs are evaluated for feature extraction, including interaction-attention, interaction-pooling, interaction-seq2seq, and seq2seq. The framework is validated through a case study of a desktop disassembly task with robots operating at different speeds. The results include evaluating different encoder designs, analyzing the impact of incorporating robot motion into the encoder, and detailed visualizations. The findings show that the proposed framework can accurately predict human trajectories and intentions.
publisherThe American Society of Mechanical Engineers (ASME)
titleMulti-Task Learning for Intention and Trajectory Prediction in Human–Robot Collaborative Disassembly Tasks
typeJournal Paper
journal volume25
journal issue5
journal titleJournal of Computing and Information Science in Engineering
identifier doi10.1115/1.4067157
journal fristpage51002-1
journal lastpage51002-12
page12
treeJournal of Computing and Information Science in Engineering:;2025:;volume( 025 ):;issue: 005
contenttypeFulltext


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record