TY - CONF TI - A Comparative Study on Deep Feature Extraction Approaches for Visual Tracking of Industrial Robots SP - 122 ID - scholars17348 KW - Convolution; Convolutional neural networks; Deep learning; Extraction; Industrial robots; Learning algorithms; Robot vision; Tracking (position) KW - Comparatives studies; Convolutional neural network; Deep learning; Feature extraction methods; Feature learning; Feature representation; Features extraction; Transformer network; Visual object tracking; Visual Tracking KW - Feature extraction N2 - Owing to several breakthroughs in deep-learning research, visual object trackers based on deep learning algorithms have been the de facto method in the tracking community. However, the performance of this category of trackers is largely influenced by the richness and quality of the target feature representation. Thus, to extract features with high discriminative power, special attention must be paid to the feature extraction component of the tracker. Despite the significance of feature extraction paradigm in deep learning-based trackers, limited studies have been dedicated to comparatively examining the feature learning and representation capabilities of various existing deep feature extraction approaches in visual tracking. In this paper, we present a comparative study on two different deep feature extraction methods in visual tracking of industrial robots in video. This work examines and visualizes the feature representation capability of two commonly employed feature extraction methods in deep learning-based tracking; namely, Convolutional Neural Network (CNN)-based and Transformer-based feature extraction methods. To demonstrate the effectiveness of the deep learning-based tracking paradigm, this study adopts a Siamese-based tracker that is trained end-To-end offline with ILSVRC15 datasets as well as a Transformer-based tracking architecture that is also pre-Trained on the ImageNet dataset. This study also contributes a dataset consisting of images of industrial robots, which was employed for evaluating the two adopted trackers. From this study, we could conclude that the feature learning and representation of the transformer-based feature extraction pipeline largely improved the overall tracking performance of the tracker. © 2022 IEEE. N1 - cited By 1; Conference of 3rd IEEE Industrial Electronics and Applications Conference, IEACon 2022 ; Conference Date: 3 October 2022 Through 4 October 2022; Conference Code:184531 AV - none EP - 127 A1 - Amosa, T.I. A1 - Sebastian, P. A1 - Izhar, L.I.B. A1 - Ibrahim, O. UR - https://www.scopus.com/inward/record.uri?eid=2-s2.0-85143382815&doi=10.1109%2fIEACon55029.2022.9951767&partnerID=40&md5=250d399cb7e6b1474a13394582086b61 PB - Institute of Electrical and Electronics Engineers Inc. SN - 9781665480062 Y1 - 2022/// ER -