Multi-Task Learning for Intention and Trajectory Prediction in Human–Robot Collaborative Disassembly TasksSource: Journal of Computing and Information Science in Engineering:;2025:;volume( 025 ):;issue: 005::page 51002-1DOI: 10.1115/1.4067157Publisher: The American Society of Mechanical Engineers (ASME)
Abstract: Human–robot collaboration (HRC) has become an integral element of many manufacturing and service industries. A fundamental requirement for safe HRC is understanding and predicting human trajectories and intentions, especially when humans and robots operate nearby. Although existing research emphasizes predicting human motions or intentions, a key challenge is predicting both human trajectories and intentions simultaneously. This paper addresses this gap by developing a multi-task learning framework consisting of a bi-long short-term memory-based encoder–decoder architecture that obtains the motion data from both human and robot trajectories as inputs and performs two main tasks simultaneously: human trajectory prediction and human intention prediction. The first task predicts human trajectories by reconstructing the motion sequences, while the second task tests two main approaches for intention prediction: supervised learning, specifically a support vector machine, to predict human intention based on the latent representation, and, an unsupervised learning method, the hidden Markov model, that decodes the latent features for human intention prediction. Four encoder designs are evaluated for feature extraction, including interaction-attention, interaction-pooling, interaction-seq2seq, and seq2seq. The framework is validated through a case study of a desktop disassembly task with robots operating at different speeds. The results include evaluating different encoder designs, analyzing the impact of incorporating robot motion into the encoder, and detailed visualizations. The findings show that the proposed framework can accurately predict human trajectories and intentions.
|
Show full item record
contributor author | Zhang, Xinyao | |
contributor author | Tian, Sibo | |
contributor author | Liang, Xiao | |
contributor author | Zheng, Minghui | |
contributor author | Behdad, Sara | |
date accessioned | 2025-08-20T09:30:28Z | |
date available | 2025-08-20T09:30:28Z | |
date copyright | 3/10/2025 12:00:00 AM | |
date issued | 2025 | |
identifier issn | 1530-9827 | |
identifier other | jcise-24-1293.pdf | |
identifier uri | http://yetl.yabesh.ir/yetl1/handle/yetl/4308389 | |
description abstract | Human–robot collaboration (HRC) has become an integral element of many manufacturing and service industries. A fundamental requirement for safe HRC is understanding and predicting human trajectories and intentions, especially when humans and robots operate nearby. Although existing research emphasizes predicting human motions or intentions, a key challenge is predicting both human trajectories and intentions simultaneously. This paper addresses this gap by developing a multi-task learning framework consisting of a bi-long short-term memory-based encoder–decoder architecture that obtains the motion data from both human and robot trajectories as inputs and performs two main tasks simultaneously: human trajectory prediction and human intention prediction. The first task predicts human trajectories by reconstructing the motion sequences, while the second task tests two main approaches for intention prediction: supervised learning, specifically a support vector machine, to predict human intention based on the latent representation, and, an unsupervised learning method, the hidden Markov model, that decodes the latent features for human intention prediction. Four encoder designs are evaluated for feature extraction, including interaction-attention, interaction-pooling, interaction-seq2seq, and seq2seq. The framework is validated through a case study of a desktop disassembly task with robots operating at different speeds. The results include evaluating different encoder designs, analyzing the impact of incorporating robot motion into the encoder, and detailed visualizations. The findings show that the proposed framework can accurately predict human trajectories and intentions. | |
publisher | The American Society of Mechanical Engineers (ASME) | |
title | Multi-Task Learning for Intention and Trajectory Prediction in Human–Robot Collaborative Disassembly Tasks | |
type | Journal Paper | |
journal volume | 25 | |
journal issue | 5 | |
journal title | Journal of Computing and Information Science in Engineering | |
identifier doi | 10.1115/1.4067157 | |
journal fristpage | 51002-1 | |
journal lastpage | 51002-12 | |
page | 12 | |
tree | Journal of Computing and Information Science in Engineering:;2025:;volume( 025 ):;issue: 005 | |
contenttype | Fulltext |