eprintid: 17348 rev_number: 2 eprint_status: archive userid: 1 dir: disk0/00/01/73/48 datestamp: 2023-12-19 03:23:45 lastmod: 2023-12-19 03:23:45 status_changed: 2023-12-19 03:07:54 type: conference_item metadata_visibility: show creators_name: Amosa, T.I. creators_name: Sebastian, P. creators_name: Izhar, L.I.B. creators_name: Ibrahim, O. title: A Comparative Study on Deep Feature Extraction Approaches for Visual Tracking of Industrial Robots ispublished: pub keywords: Convolution; Convolutional neural networks; Deep learning; Extraction; Industrial robots; Learning algorithms; Robot vision; Tracking (position), Comparatives studies; Convolutional neural network; Deep learning; Feature extraction methods; Feature learning; Feature representation; Features extraction; Transformer network; Visual object tracking; Visual Tracking, Feature extraction note: cited By 1; Conference of 3rd IEEE Industrial Electronics and Applications Conference, IEACon 2022 ; Conference Date: 3 October 2022 Through 4 October 2022; Conference Code:184531 abstract: Owing to several breakthroughs in deep-learning research, visual object trackers based on deep learning algorithms have been the de facto method in the tracking community. However, the performance of this category of trackers is largely influenced by the richness and quality of the target feature representation. Thus, to extract features with high discriminative power, special attention must be paid to the feature extraction component of the tracker. Despite the significance of feature extraction paradigm in deep learning-based trackers, limited studies have been dedicated to comparatively examining the feature learning and representation capabilities of various existing deep feature extraction approaches in visual tracking. In this paper, we present a comparative study on two different deep feature extraction methods in visual tracking of industrial robots in video. This work examines and visualizes the feature representation capability of two commonly employed feature extraction methods in deep learning-based tracking; namely, Convolutional Neural Network (CNN)-based and Transformer-based feature extraction methods. To demonstrate the effectiveness of the deep learning-based tracking paradigm, this study adopts a Siamese-based tracker that is trained end-To-end offline with ILSVRC15 datasets as well as a Transformer-based tracking architecture that is also pre-Trained on the ImageNet dataset. This study also contributes a dataset consisting of images of industrial robots, which was employed for evaluating the two adopted trackers. From this study, we could conclude that the feature learning and representation of the transformer-based feature extraction pipeline largely improved the overall tracking performance of the tracker. © 2022 IEEE. date: 2022 publisher: Institute of Electrical and Electronics Engineers Inc. official_url: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85143382815&doi=10.1109%2fIEACon55029.2022.9951767&partnerID=40&md5=250d399cb7e6b1474a13394582086b61 id_number: 10.1109/IEACon55029.2022.9951767 full_text_status: none publication: 2022 IEEE Industrial Electronics and Applications Conference, IEACon 2022 pagerange: 122-127 refereed: TRUE isbn: 9781665480062 citation: Amosa, T.I. and Sebastian, P. and Izhar, L.I.B. and Ibrahim, O. (2022) A Comparative Study on Deep Feature Extraction Approaches for Visual Tracking of Industrial Robots. In: UNSPECIFIED.