Article Search
닫기

Original Article

Split Viewer

International Journal of Fuzzy Logic and Intelligent Systems 2024; 24(3): 242-257

Published online September 25, 2024

https://doi.org/10.5391/IJFIS.2024.24.3.242

© The Korean Institute of Intelligent Systems

Motion Imitation Robot Based on Artificial Neural Network with Minimization of Restrictions on Degrees of Freedom

Jeong-Hun Kang1, Seong-Jin Park2, Ye-Won Kim1, and Bo-Yeong Kang3

1Department of Artificial Intelligence, Kyungpook National University, Daegu, Korea
2Department of Mechanical Engineering, Kyungpook National University, Daegu, Korea
3Department of Robot and Smart System Engineering, Kyungpook National University, Daegu, Korea

Correspondence to :
Bo-Yeong Kang (kby09@knu.ac.kr)

Received: June 16, 2022; Revised: May 18, 2024; Accepted: September 20, 2024

This is an Open Access article distributed under the terms of the Creative Commons Attribution Non-Commercial License (http://creativecommons.org/licenses/by-nc/3.0/) which permits unrestricted noncommercial use, distribution, and reproduction in any medium, provided the original work is properly cited.

For a robot to imitate human motions, each human joint must be mapped onto the robot. In the mapping process of the NAO robot, there is a degrees-of-freedom mismatch problem between a human arm with six degrees of freedom and a robot arm with four degrees of freedom. During the collection of information on robot joint angles from human joint angles, some information on the six degrees of freedom is absent, resulting in inaccurate or erroneous movements of the robot, requiring additional calculations. In this paper, we propose a robot technology that imitates human movements by minimizing the degrees-of-freedom constraints without missing information using an artificial neural network. To verify the proposed approach, a manually measured answer dataset and an inverse kinematics answer dataset were created for each of the 919 motion frames of the human right-arm and upper-body motions. The robot imitation performance was stable through a 10-fold verification with the manually measured and inverse kinematics answer datasets for the right-arm motion imitations of 3.245 and 4.24 and the upper-body imitations of 5.10 and 4.82. In addition, as the trends of the robot prediction motion signal graph were similar to those of the answer motion signal graph, the proposed approach demonstrated a steady imitation performance.

Keywords: Artificial neural network, Motion imitation, Artificial intelligence (AI), NAO robot

This study was supported by the National Research Foundation of Korea and funded by the Korean Government under Grant NRF-2019R1A2C1011270.

No potential conflict of interest relevant to this article was reported.

Jeong-Hun Kang received his B.S. degree from the Department of Precision Mechanical Engineering and School of Computer Science at Kyungpook National University, Daegu, Korea in 2020. He is currently pursuing his master’s degree from the Department of Artificial Intelligence. His research interests include reinforcement learning, deep learning, and cooperative robotics.

E-mail: jhkang.knu@gmail.com

Seong-Jin Park received his B.S. degree from the Department of Mechanical Engineering at Kyungpook National University, Daegu, Korea, in 2021.

E-mail: qwerty78766@naver.com

Ye-Won Kim received her B.S. and M.S. degrees from Kyungpook National University, Daegu, Korea in 2017 and 2020, respectively. She is currently pursuing her Ph.D. from the Department of Artificial Intelligence. Her research interests include reinforcement learning, deep learning, and cooperative robotics.

E-mail: yewonkim.knu@gmail.com

Bo-Yeong Kang, received her M.S. and Ph.D. degrees in computer engineering from Kyungpook National University, Daegu, Korea in 2002 and 2004, respectively. She is currently a professor in the Department of Robot and Smart Systems Engineering. Her research interests include reinforcement learning, deep learning, cooperative robots, and artificial intelligence.

E-mail: kby09@knu.ac.kr

Article

Original Article

International Journal of Fuzzy Logic and Intelligent Systems 2024; 24(3): 242-257

Published online September 25, 2024 https://doi.org/10.5391/IJFIS.2024.24.3.242

Copyright © The Korean Institute of Intelligent Systems.

Motion Imitation Robot Based on Artificial Neural Network with Minimization of Restrictions on Degrees of Freedom

Jeong-Hun Kang1, Seong-Jin Park2, Ye-Won Kim1, and Bo-Yeong Kang3

1Department of Artificial Intelligence, Kyungpook National University, Daegu, Korea
2Department of Mechanical Engineering, Kyungpook National University, Daegu, Korea
3Department of Robot and Smart System Engineering, Kyungpook National University, Daegu, Korea

Correspondence to:Bo-Yeong Kang (kby09@knu.ac.kr)

Received: June 16, 2022; Revised: May 18, 2024; Accepted: September 20, 2024

This is an Open Access article distributed under the terms of the Creative Commons Attribution Non-Commercial License (http://creativecommons.org/licenses/by-nc/3.0/) which permits unrestricted noncommercial use, distribution, and reproduction in any medium, provided the original work is properly cited.

Abstract

For a robot to imitate human motions, each human joint must be mapped onto the robot. In the mapping process of the NAO robot, there is a degrees-of-freedom mismatch problem between a human arm with six degrees of freedom and a robot arm with four degrees of freedom. During the collection of information on robot joint angles from human joint angles, some information on the six degrees of freedom is absent, resulting in inaccurate or erroneous movements of the robot, requiring additional calculations. In this paper, we propose a robot technology that imitates human movements by minimizing the degrees-of-freedom constraints without missing information using an artificial neural network. To verify the proposed approach, a manually measured answer dataset and an inverse kinematics answer dataset were created for each of the 919 motion frames of the human right-arm and upper-body motions. The robot imitation performance was stable through a 10-fold verification with the manually measured and inverse kinematics answer datasets for the right-arm motion imitations of 3.245 and 4.24 and the upper-body imitations of 5.10 and 4.82. In addition, as the trends of the robot prediction motion signal graph were similar to those of the answer motion signal graph, the proposed approach demonstrated a steady imitation performance.

Keywords: Artificial neural network, Motion imitation, Artificial intelligence (AI), NAO robot

Fig 1.

Figure 1.

Robot motion prediction work diagram of the proposed technology for the imitation of human motion of robots and ANN construction.

The International Journal of Fuzzy Logic and Intelligent Systems 2024; 24: 242-257https://doi.org/10.5391/IJFIS.2024.24.3.242

Fig 2.

Figure 2.

Result loss graph of each motion with a manually measured dataset and an inverse kinematics dataset with the ANN: (a–c) inverse kinematic loss graph (walking-jumping-greeting) and (d–f) manually measured loss graph (walking-jumping-greeting).

The International Journal of Fuzzy Logic and Intelligent Systems 2024; 24: 242-257https://doi.org/10.5391/IJFIS.2024.24.3.242

Fig 3.

Figure 3.

Dataset configuration.

The International Journal of Fuzzy Logic and Intelligent Systems 2024; 24: 242-257https://doi.org/10.5391/IJFIS.2024.24.3.242

Fig 4.

Figure 4.

Robot imitation of human right-arm motion (left, human motion; right, robot motion) and result graph of each motion with the manual measurement dataset and inverse kinematics dataset: (a–c) right-arm motion shape (walking-jumping-greeting), (d–f) right-arm signal answer with inverse kinematics, (g–i) right-arm signal prediction with inverse kinematics, (j–l) right-arm signal answer with manual measurement, and (m–o) right-arm signal prediction with manual measurement. The x-axis represents time, and the y-axis represents the angle.

The International Journal of Fuzzy Logic and Intelligent Systems 2024; 24: 242-257https://doi.org/10.5391/IJFIS.2024.24.3.242

Fig 5.

Figure 5.

Robot imitation of human upper-body motion (left, human motion; right, robot motion) and result graph of each motion with the manual measurement dataset and inverse kinematics dataset: (a–c) upper-body motion shape (walking-jumping-greeting), (d–f) upper-body signal answer with inverse kinematics, (g–i) upper-body signal prediction with inverse kinematics, (j–l) upper-body signal answer with manual measurement, and (m–o) upper-body signal prediction with manual measurement. The x-axis represents time, and the y-axis represents the angle.

The International Journal of Fuzzy Logic and Intelligent Systems 2024; 24: 242-257https://doi.org/10.5391/IJFIS.2024.24.3.242

Fig 6.

Figure 6.

Figure A1. Result graph of each motion with the manual measurement dataset and inverse kinematics dataset showing the difference between the answer and predicted angle values: (a–c) difference in the joint angles of the right arm with inverse kinematics, (d–f) difference in the joint angles of the right arm with manual measurement, (g–i) difference in the joint angles of the upper body with inverse kinematics, and (j–l) difference in the joint angles of the upper body with manual measurement. The x-axis represents time, and the y-axis represents the angle.

The International Journal of Fuzzy Logic and Intelligent Systems 2024; 24: 242-257https://doi.org/10.5391/IJFIS.2024.24.3.242

Fig 7.

Figure 7.

Figure A2. Robot imitation of human right-arm motion (left, human motion; right, robot motion) and result graph of each motion with the manual measurement dataset and inverse kinematics dataset, trained for 5,000 epochs: (a–c) right-armmotion shape (walking-jumping-greeting), (d–f) right-arm signal answer with inverse kinematics, (g–i) right-arm signal prediction with inverse kinematics, (j–l) right-arm signal answer with manual measurement, and (m–o) right-arm signal prediction with manual measurement. The x-axis represents time, and the y-axis represents the angle.

The International Journal of Fuzzy Logic and Intelligent Systems 2024; 24: 242-257https://doi.org/10.5391/IJFIS.2024.24.3.242

Fig 8.

Figure 8.

Figure A3. Robot imitation of human right-arm motion (left, human motion; right, robot motion) and result graph of each motion with the manual measurement dataset and inverse kinematics dataset, trained for 15,000 epochs: (a–c) right-arm motion shape (walking-jumping-greeting), (d–f) right-arm signal answer with inverse kinematics, (g–i) right-arm signal prediction with inverse kinematics, (j–l) right-arm signal answer with manual measurement, and (m–o) right-arm signal prediction with manual measurement. The x-axis represents time, and the y-axis represents the angle.

The International Journal of Fuzzy Logic and Intelligent Systems 2024; 24: 242-257https://doi.org/10.5391/IJFIS.2024.24.3.242

Fig 9.

Figure 9.

Figure A4. Robot imitation of human upper-body motion (left, human motion; right, robot motion) and result graph of each motion with the manual measurement dataset and inverse kinematics dataset, trained for 5,000 epochs: (a–c) upper-body motion shape (walking-jumping-greeting), (d–f) upper-body signal answer with inverse kinematics, (g–i) upper-body signal prediction with inverse kinematics, (j–l) upper-body signal answer with manual measurement, and (m–o) upper-body signal prediction with manual measurement. The x-axis represents time, and the y-axis represents the angle.

The International Journal of Fuzzy Logic and Intelligent Systems 2024; 24: 242-257https://doi.org/10.5391/IJFIS.2024.24.3.242

Fig 10.

Figure 10.

Figure A5. Robot imitation of human upper-body motion (left, human motion; right, robot motion) and result graph of each motion with the manual measurement dataset and inverse kinematics dataset, trained for 15,000 epochs: (a–c) upper-body motion shape (walking-jumping-greeting), (d–f) upper-body signal answer with inverse kinematics, (g–i) upper-body signal prediction with inverse kinematics, (j–l) upper-body signal answer with manual measurement, and (m–o) upper-body signal prediction with manual measurement. The x-axis represents time, and the y-axis represents the angle.

The International Journal of Fuzzy Logic and Intelligent Systems 2024; 24: 242-257https://doi.org/10.5391/IJFIS.2024.24.3.242

Table 1 . Average error of 10-fold cross-validation and test set validation for right-arm motion imitation (unit: degree).

10-fold testTest set
Manually measured_Proposed3.2450 ± 0.52.9743 ± 0.5
Inverse kinematics_Proposed4.2357 ± 0.54.6399 ± 0.5

Table 2 . Average error of 10-fold cross-validation and test set validation for upper-body motion imitation (unit: degree).

10-fold testTest set
Manually measured_Proposed5.1018 ± 0.55.3845 ± 0.5
Inverse kinematics_Proposed4.8150 ± 0.54.6146 ± 0.5

Share this article on :

Related articles in IJFIS