DSpace Repository

DEEP LEARNING-BASED OBJECT CLASSIFICATION AND POSITION ESTIMATION PIPELINE FOR POTENTIAL USE IN ROBOTIZED PICK-AND-PLACE OPERATIONS

Система будет остановлена для регулярного обслуживания. Пожалуйста, сохраните рабочие данные и выйдите из системы.

Show simple item record

dc.contributor.author Soltan, Sergey
dc.contributor.author Oleinikov, Artemiy
dc.contributor.author Demirci, M. Fatih
dc.contributor.author Shintemirov, Almas
dc.date.accessioned 2021-07-02T07:07:25Z
dc.date.available 2021-07-02T07:07:25Z
dc.date.issued 2020-08
dc.identifier.citation Soltan, S., Oleinikov, A., Demirci, M. F., & Shintemirov, A. (2020). Deep Learning-Based Object Classification and Position Estimation Pipeline for Potential Use in Robotized Pick-and-Place Operations. Robotics, 9(3), 63. https://doi.org/10.3390/robotics9030063 en_US
dc.identifier.uri http://nur.nu.edu.kz/handle/123456789/5515
dc.description.abstract Accurate object classification and position estimation is a crucial part of executing autonomous pick-and-place operations by a robot and can be realized using RGB-D sensors becoming increasingly available for use in industrial applications. In this paper, we present a novel unified framework for object detection and classification using a combination of point cloud processing and deep learning techniques. The proposed model uses two streams that recognize objects on RGB and depth data separately and combines the two in later stages to classify objects. Experimental evaluation of the proposed model including classification accuracy compared with previous works demonstrates its effectiveness and efficiency, making the model suitable for real-time applications. In particular, the experiments performed on the Washington RGB-D object dataset show that the proposed framework has 97.5% and 95% fewer parameters compared to the previous state-of-the-art multimodel neural networks Fus-CNN, CNN Features and VGG3D, respectively, with the cost of approximately 5% drop in classification accuracy. Moreover, the inference of the proposed framework takes 66.11%, 32.65%, and 28.77% less time on GPU and 86.91%, 51.12%, and 50.15% less time on CPU in comparison to VGG3D, Fus-CNN, and CNN Features. The potential applicability of the developed object classification and position estimation framework was then demonstrated on an experimental robot-manipulation setup realizing a simplified object pick-and-place scenario. In approximately 95% of test trials, the system was able to accurately position the robot over the detected objects of interest in an automatic mode, ensuring stable cyclic execution with no time delays en_US
dc.language.iso en en_US
dc.publisher Robotics en_US
dc.rights Attribution-NonCommercial-ShareAlike 3.0 United States *
dc.rights.uri http://creativecommons.org/licenses/by-nc-sa/3.0/us/ *
dc.subject object classification en_US
dc.subject object position estimation en_US
dc.subject deep learning en_US
dc.subject neural network en_US
dc.title DEEP LEARNING-BASED OBJECT CLASSIFICATION AND POSITION ESTIMATION PIPELINE FOR POTENTIAL USE IN ROBOTIZED PICK-AND-PLACE OPERATIONS en_US
dc.type Article en_US
workflow.import.source science


Files in this item

The following license files are associated with this item:

This item appears in the following Collection(s)

Show simple item record

Attribution-NonCommercial-ShareAlike 3.0 United States Except where otherwise noted, this item's license is described as Attribution-NonCommercial-ShareAlike 3.0 United States