Best Frame Selection to Enhance Training Step Efficiency in Video-Based Human Action Recognition
Autor(a) principal: | |
---|---|
Data de Publicação: | 2022 |
Outros Autores: | , , , |
Tipo de documento: | Artigo |
Idioma: | eng |
Título da fonte: | Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos) |
Texto Completo: | https://hdl.handle.net/10216/139959 |
Resumo: | In recent years, with the growth of digital media and modern imaging equipment, the use of video processing algorithms and semantic film and image management has expanded. The usage of different video datasets in training artificial intelligence algorithms is also rapidly expanding in various fields. Due to the high volume of information in a video, its processing is still expensive for most hardware systems, mainly in terms of its required runtime and memory. Hence, the optimal selection of keyframes to minimize redundant information in video processing systems has become noteworthy in facilitating this problem. Eliminating some frames can simultaneously reduce the required computational load, hardware cost, memory and processing time of intelligent video-based systems. Based on the aforementioned reasons, this research proposes a method for selecting keyframes and adaptive cropping input video for human action recognition (HAR) systems. The proposed method combines edge detection, simple difference, adaptive thresholding and 1D and 2D average filter algorithms in a hierarchical method. Some HAR methods are trained with videos processed by the proposed method to assess its efficiency. The results demonstrate that the application of the proposed method increases the accuracy of the HAR system by up to 3% compared to random image selection and cropping methods. Additionally, for most cases, the proposed method reduces the training time of the used machine learning algorithm. |
id |
RCAP_b86f502ca1fc41476d7bd7acdaf7e839 |
---|---|
oai_identifier_str |
oai:repositorio-aberto.up.pt:10216/139959 |
network_acronym_str |
RCAP |
network_name_str |
Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos) |
repository_id_str |
7160 |
spelling |
Best Frame Selection to Enhance Training Step Efficiency in Video-Based Human Action RecognitionCiências Tecnológicas, Ciências da engenharia e tecnologiasTechnological sciences, Engineering and technologyIn recent years, with the growth of digital media and modern imaging equipment, the use of video processing algorithms and semantic film and image management has expanded. The usage of different video datasets in training artificial intelligence algorithms is also rapidly expanding in various fields. Due to the high volume of information in a video, its processing is still expensive for most hardware systems, mainly in terms of its required runtime and memory. Hence, the optimal selection of keyframes to minimize redundant information in video processing systems has become noteworthy in facilitating this problem. Eliminating some frames can simultaneously reduce the required computational load, hardware cost, memory and processing time of intelligent video-based systems. Based on the aforementioned reasons, this research proposes a method for selecting keyframes and adaptive cropping input video for human action recognition (HAR) systems. The proposed method combines edge detection, simple difference, adaptive thresholding and 1D and 2D average filter algorithms in a hierarchical method. Some HAR methods are trained with videos processed by the proposed method to assess its efficiency. The results demonstrate that the application of the proposed method increases the accuracy of the HAR system by up to 3% compared to random image selection and cropping methods. Additionally, for most cases, the proposed method reduces the training time of the used machine learning algorithm.2022-022022-02-01T00:00:00Zinfo:eu-repo/semantics/publishedVersioninfo:eu-repo/semantics/articleapplication/pdfimage/jpeghttps://hdl.handle.net/10216/139959eng10.3390/app12041830Abdorreza Alavi GharahbaghVahid HajihashemiMarta Campos FerreiraJosé J. M. MachadoJoão Manuel R. S. Tavaresinfo:eu-repo/semantics/openAccessreponame:Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos)instname:Agência para a Sociedade do Conhecimento (UMIC) - FCT - Sociedade da Informaçãoinstacron:RCAAP2023-11-29T12:42:14Zoai:repositorio-aberto.up.pt:10216/139959Portal AgregadorONGhttps://www.rcaap.pt/oai/openaireopendoar:71602024-03-19T23:25:05.776471Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos) - Agência para a Sociedade do Conhecimento (UMIC) - FCT - Sociedade da Informaçãofalse |
dc.title.none.fl_str_mv |
Best Frame Selection to Enhance Training Step Efficiency in Video-Based Human Action Recognition |
title |
Best Frame Selection to Enhance Training Step Efficiency in Video-Based Human Action Recognition |
spellingShingle |
Best Frame Selection to Enhance Training Step Efficiency in Video-Based Human Action Recognition Abdorreza Alavi Gharahbagh Ciências Tecnológicas, Ciências da engenharia e tecnologias Technological sciences, Engineering and technology |
title_short |
Best Frame Selection to Enhance Training Step Efficiency in Video-Based Human Action Recognition |
title_full |
Best Frame Selection to Enhance Training Step Efficiency in Video-Based Human Action Recognition |
title_fullStr |
Best Frame Selection to Enhance Training Step Efficiency in Video-Based Human Action Recognition |
title_full_unstemmed |
Best Frame Selection to Enhance Training Step Efficiency in Video-Based Human Action Recognition |
title_sort |
Best Frame Selection to Enhance Training Step Efficiency in Video-Based Human Action Recognition |
author |
Abdorreza Alavi Gharahbagh |
author_facet |
Abdorreza Alavi Gharahbagh Vahid Hajihashemi Marta Campos Ferreira José J. M. Machado João Manuel R. S. Tavares |
author_role |
author |
author2 |
Vahid Hajihashemi Marta Campos Ferreira José J. M. Machado João Manuel R. S. Tavares |
author2_role |
author author author author |
dc.contributor.author.fl_str_mv |
Abdorreza Alavi Gharahbagh Vahid Hajihashemi Marta Campos Ferreira José J. M. Machado João Manuel R. S. Tavares |
dc.subject.por.fl_str_mv |
Ciências Tecnológicas, Ciências da engenharia e tecnologias Technological sciences, Engineering and technology |
topic |
Ciências Tecnológicas, Ciências da engenharia e tecnologias Technological sciences, Engineering and technology |
description |
In recent years, with the growth of digital media and modern imaging equipment, the use of video processing algorithms and semantic film and image management has expanded. The usage of different video datasets in training artificial intelligence algorithms is also rapidly expanding in various fields. Due to the high volume of information in a video, its processing is still expensive for most hardware systems, mainly in terms of its required runtime and memory. Hence, the optimal selection of keyframes to minimize redundant information in video processing systems has become noteworthy in facilitating this problem. Eliminating some frames can simultaneously reduce the required computational load, hardware cost, memory and processing time of intelligent video-based systems. Based on the aforementioned reasons, this research proposes a method for selecting keyframes and adaptive cropping input video for human action recognition (HAR) systems. The proposed method combines edge detection, simple difference, adaptive thresholding and 1D and 2D average filter algorithms in a hierarchical method. Some HAR methods are trained with videos processed by the proposed method to assess its efficiency. The results demonstrate that the application of the proposed method increases the accuracy of the HAR system by up to 3% compared to random image selection and cropping methods. Additionally, for most cases, the proposed method reduces the training time of the used machine learning algorithm. |
publishDate |
2022 |
dc.date.none.fl_str_mv |
2022-02 2022-02-01T00:00:00Z |
dc.type.status.fl_str_mv |
info:eu-repo/semantics/publishedVersion |
dc.type.driver.fl_str_mv |
info:eu-repo/semantics/article |
format |
article |
status_str |
publishedVersion |
dc.identifier.uri.fl_str_mv |
https://hdl.handle.net/10216/139959 |
url |
https://hdl.handle.net/10216/139959 |
dc.language.iso.fl_str_mv |
eng |
language |
eng |
dc.relation.none.fl_str_mv |
10.3390/app12041830 |
dc.rights.driver.fl_str_mv |
info:eu-repo/semantics/openAccess |
eu_rights_str_mv |
openAccess |
dc.format.none.fl_str_mv |
application/pdf image/jpeg |
dc.source.none.fl_str_mv |
reponame:Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos) instname:Agência para a Sociedade do Conhecimento (UMIC) - FCT - Sociedade da Informação instacron:RCAAP |
instname_str |
Agência para a Sociedade do Conhecimento (UMIC) - FCT - Sociedade da Informação |
instacron_str |
RCAAP |
institution |
RCAAP |
reponame_str |
Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos) |
collection |
Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos) |
repository.name.fl_str_mv |
Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos) - Agência para a Sociedade do Conhecimento (UMIC) - FCT - Sociedade da Informação |
repository.mail.fl_str_mv |
|
_version_ |
1799135555960700928 |