Yayın:
Validating assembly tasks in real time using deep learning for object detection and tracking the operator’s hand-joint movements.

dc.contributor.authorAy, Öznur
dc.contributor.authorEmel, Erdal
dc.contributor.buuauthorAy, Öznur
dc.contributor.buuauthorEMEL, ERDAL
dc.contributor.departmentMühendislik Fakültesi
dc.contributor.departmentEndüstri Mühendisliği Bölümü
dc.contributor.orcid0000-0002-9220-7353
dc.contributor.scopusid59712710000
dc.contributor.scopusid6602919521
dc.date.accessioned2025-11-28T12:15:40Z
dc.date.issued2025-01-01
dc.description.abstractReal-time monitoring for evaluating manual assembly work is essential for improving operator efficiency and product quality. Recent applications detect and prevent operator errors with instant feedback by visually recognizing actions in a monitored assembly scene. However, prior studies often disregarded hand-object interactions and lacked the modeling of fine-grained hand movements. Because industrial assembly tasks are primarily performed by hand, the focus should be on the hands and their interactions with manipulated tools and objects. This paper proposes a real-time fine-grained assembly task recognition system using 3-dimensional hand skeleton data extracted from streaming 2-dimensional video frames. The hybrid task recognition system consists of a You-Only-Look-Once (YOLO) deep object detection method and a Long Short-Term Memory (LSTM) based classifier working in an integrated approach. First, from the streaming data, the possible starting point for each sequential task is determined using the object detection method. Time series skeleton data were then captured using a pose estimation algorithm from the possible starting point until YOLOv8 detected a different start point. Subsequently, the proposed LSTM-based network classifies the time series of the hand joint coordinates to comply with the corresponding fine-grained assembly task. Sequential tasks at an industrial assembly station are used to create an operator-centric video dataset with annotations to evaluate the proposed system. The proposed hybrid system significantly improved the operator efficiency for sequential assembly tasks, achieving 85.23% accuracy in real-time task recognition. The real-world industrial assembly dataset used in our study was also shared as open access for the assembly task recognition community.
dc.identifier.doi10.1109/ACCESS.2025.3554263
dc.identifier.endpage57029
dc.identifier.issn21693536
dc.identifier.scopus2-s2.0-105003088699
dc.identifier.startpage57009
dc.identifier.urihttps://hdl.handle.net/11452/57141
dc.identifier.volume13
dc.indexed.scopusScopus
dc.language.isoen
dc.publisherInstitute of Electrical and Electronics Engineers Inc.
dc.relation.bapFDK-2022-1255
dc.relation.journalIEEE Access
dc.relation.tubitak118C136
dc.rightsinfo:eu-repo/semantics/openAccess
dc.subjectTime series classification
dc.subjectPose estimation
dc.subjectObject detection
dc.subjectLong-short term memory
dc.subjectConnected worker
dc.subjectAssembly action recognition
dc.subject.scopusHuman Action Recognition Systems
dc.titleValidating assembly tasks in real time using deep learning for object detection and tracking the operator’s hand-joint movements.
dc.typeArticle
dspace.entity.typePublication
local.contributor.departmentMühendislik Fakültesi/Endüstri Mühendisliği Bölümü
local.indexed.atScopus
relation.isAuthorOfPublication758ceefe-22fa-474e-8207-c551b8f5f98a
relation.isAuthorOfPublication.latestForDiscovery758ceefe-22fa-474e-8207-c551b8f5f98a

Dosyalar

Orijinal seri

Şimdi gösteriliyor 1 - 1 / 1
Küçük Resim
Ad:
Ay_vd_2025
Boyut:
3.8 MB
Format:
Adobe Portable Document Format