Best Frame Selection to Enhance Training Step Efficiency in Video-Based Human Action Recognition

Detalhes bibliográficos
Autor(a) principal: Abdorreza Alavi Gharahbagh
Data de Publicação: 2022
Outros Autores: Vahid Hajihashemi, Marta Campos Ferreira, José J. M. Machado, João Manuel R. S. Tavares
Tipo de documento: Artigo
Idioma: eng
Título da fonte: Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos)
Texto Completo: https://hdl.handle.net/10216/139959
Resumo: In recent years, with the growth of digital media and modern imaging equipment, the use of video processing algorithms and semantic film and image management has expanded. The usage of different video datasets in training artificial intelligence algorithms is also rapidly expanding in various fields. Due to the high volume of information in a video, its processing is still expensive for most hardware systems, mainly in terms of its required runtime and memory. Hence, the optimal selection of keyframes to minimize redundant information in video processing systems has become noteworthy in facilitating this problem. Eliminating some frames can simultaneously reduce the required computational load, hardware cost, memory and processing time of intelligent video-based systems. Based on the aforementioned reasons, this research proposes a method for selecting keyframes and adaptive cropping input video for human action recognition (HAR) systems. The proposed method combines edge detection, simple difference, adaptive thresholding and 1D and 2D average filter algorithms in a hierarchical method. Some HAR methods are trained with videos processed by the proposed method to assess its efficiency. The results demonstrate that the application of the proposed method increases the accuracy of the HAR system by up to 3% compared to random image selection and cropping methods. Additionally, for most cases, the proposed method reduces the training time of the used machine learning algorithm.
id RCAP_b86f502ca1fc41476d7bd7acdaf7e839
oai_identifier_str oai:repositorio-aberto.up.pt:10216/139959
network_acronym_str RCAP
network_name_str Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos)
repository_id_str 7160
spelling Best Frame Selection to Enhance Training Step Efficiency in Video-Based Human Action RecognitionCiências Tecnológicas, Ciências da engenharia e tecnologiasTechnological sciences, Engineering and technologyIn recent years, with the growth of digital media and modern imaging equipment, the use of video processing algorithms and semantic film and image management has expanded. The usage of different video datasets in training artificial intelligence algorithms is also rapidly expanding in various fields. Due to the high volume of information in a video, its processing is still expensive for most hardware systems, mainly in terms of its required runtime and memory. Hence, the optimal selection of keyframes to minimize redundant information in video processing systems has become noteworthy in facilitating this problem. Eliminating some frames can simultaneously reduce the required computational load, hardware cost, memory and processing time of intelligent video-based systems. Based on the aforementioned reasons, this research proposes a method for selecting keyframes and adaptive cropping input video for human action recognition (HAR) systems. The proposed method combines edge detection, simple difference, adaptive thresholding and 1D and 2D average filter algorithms in a hierarchical method. Some HAR methods are trained with videos processed by the proposed method to assess its efficiency. The results demonstrate that the application of the proposed method increases the accuracy of the HAR system by up to 3% compared to random image selection and cropping methods. Additionally, for most cases, the proposed method reduces the training time of the used machine learning algorithm.2022-022022-02-01T00:00:00Zinfo:eu-repo/semantics/publishedVersioninfo:eu-repo/semantics/articleapplication/pdfimage/jpeghttps://hdl.handle.net/10216/139959eng10.3390/app12041830Abdorreza Alavi GharahbaghVahid HajihashemiMarta Campos FerreiraJosé J. M. MachadoJoão Manuel R. S. Tavaresinfo:eu-repo/semantics/openAccessreponame:Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos)instname:Agência para a Sociedade do Conhecimento (UMIC) - FCT - Sociedade da Informaçãoinstacron:RCAAP2023-11-29T12:42:14Zoai:repositorio-aberto.up.pt:10216/139959Portal AgregadorONGhttps://www.rcaap.pt/oai/openaireopendoar:71602024-03-19T23:25:05.776471Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos) - Agência para a Sociedade do Conhecimento (UMIC) - FCT - Sociedade da Informaçãofalse
dc.title.none.fl_str_mv Best Frame Selection to Enhance Training Step Efficiency in Video-Based Human Action Recognition
title Best Frame Selection to Enhance Training Step Efficiency in Video-Based Human Action Recognition
spellingShingle Best Frame Selection to Enhance Training Step Efficiency in Video-Based Human Action Recognition
Abdorreza Alavi Gharahbagh
Ciências Tecnológicas, Ciências da engenharia e tecnologias
Technological sciences, Engineering and technology
title_short Best Frame Selection to Enhance Training Step Efficiency in Video-Based Human Action Recognition
title_full Best Frame Selection to Enhance Training Step Efficiency in Video-Based Human Action Recognition
title_fullStr Best Frame Selection to Enhance Training Step Efficiency in Video-Based Human Action Recognition
title_full_unstemmed Best Frame Selection to Enhance Training Step Efficiency in Video-Based Human Action Recognition
title_sort Best Frame Selection to Enhance Training Step Efficiency in Video-Based Human Action Recognition
author Abdorreza Alavi Gharahbagh
author_facet Abdorreza Alavi Gharahbagh
Vahid Hajihashemi
Marta Campos Ferreira
José J. M. Machado
João Manuel R. S. Tavares
author_role author
author2 Vahid Hajihashemi
Marta Campos Ferreira
José J. M. Machado
João Manuel R. S. Tavares
author2_role author
author
author
author
dc.contributor.author.fl_str_mv Abdorreza Alavi Gharahbagh
Vahid Hajihashemi
Marta Campos Ferreira
José J. M. Machado
João Manuel R. S. Tavares
dc.subject.por.fl_str_mv Ciências Tecnológicas, Ciências da engenharia e tecnologias
Technological sciences, Engineering and technology
topic Ciências Tecnológicas, Ciências da engenharia e tecnologias
Technological sciences, Engineering and technology
description In recent years, with the growth of digital media and modern imaging equipment, the use of video processing algorithms and semantic film and image management has expanded. The usage of different video datasets in training artificial intelligence algorithms is also rapidly expanding in various fields. Due to the high volume of information in a video, its processing is still expensive for most hardware systems, mainly in terms of its required runtime and memory. Hence, the optimal selection of keyframes to minimize redundant information in video processing systems has become noteworthy in facilitating this problem. Eliminating some frames can simultaneously reduce the required computational load, hardware cost, memory and processing time of intelligent video-based systems. Based on the aforementioned reasons, this research proposes a method for selecting keyframes and adaptive cropping input video for human action recognition (HAR) systems. The proposed method combines edge detection, simple difference, adaptive thresholding and 1D and 2D average filter algorithms in a hierarchical method. Some HAR methods are trained with videos processed by the proposed method to assess its efficiency. The results demonstrate that the application of the proposed method increases the accuracy of the HAR system by up to 3% compared to random image selection and cropping methods. Additionally, for most cases, the proposed method reduces the training time of the used machine learning algorithm.
publishDate 2022
dc.date.none.fl_str_mv 2022-02
2022-02-01T00:00:00Z
dc.type.status.fl_str_mv info:eu-repo/semantics/publishedVersion
dc.type.driver.fl_str_mv info:eu-repo/semantics/article
format article
status_str publishedVersion
dc.identifier.uri.fl_str_mv https://hdl.handle.net/10216/139959
url https://hdl.handle.net/10216/139959
dc.language.iso.fl_str_mv eng
language eng
dc.relation.none.fl_str_mv 10.3390/app12041830
dc.rights.driver.fl_str_mv info:eu-repo/semantics/openAccess
eu_rights_str_mv openAccess
dc.format.none.fl_str_mv application/pdf
image/jpeg
dc.source.none.fl_str_mv reponame:Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos)
instname:Agência para a Sociedade do Conhecimento (UMIC) - FCT - Sociedade da Informação
instacron:RCAAP
instname_str Agência para a Sociedade do Conhecimento (UMIC) - FCT - Sociedade da Informação
instacron_str RCAAP
institution RCAAP
reponame_str Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos)
collection Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos)
repository.name.fl_str_mv Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos) - Agência para a Sociedade do Conhecimento (UMIC) - FCT - Sociedade da Informação
repository.mail.fl_str_mv
_version_ 1799135555960700928