Tags:Computer vision, Deep Learning, Robotics, System Identification and Transfer Learning
Abstract:
Pose estimation from high-dimensional inputs, such as video input is a challenging system identification problem. In robotics an accurate model can provide better performance in model-predictive control (MPC). In this work, the identification input is the video of an original robotic manipulator link with a flexible joint and the position of the link the output. The methodology is implemented using deep convolutional neural networks (ConvNet). Transfer Learning (TL) approach is used, in order to optimize the training. Five known ConvNets classification models are used. The estimation made by the deep TL ConvNets and the measured angular position from the original assembly. The TL method shows an important ability to estimate the angular position of the link from the frames of the video, with every model obtaining over 0.98 determinant coefficient score. The VGG19 model shows the best accuracy at estimating the link’s position from the frames of the video, decreasing the Mean Absolute Error (MAE) by 25.71% compared to the second-best model.
Pose Estimation of Robotic Manipulators Using Deep Transfer Learning Towards Video-Based System Identification