Aircraft Maintenance Check Scheduling Using Reinforcement Learning
Autor(a) principal: | |
---|---|
Data de Publicação: | 2021 |
Outros Autores: | , , |
Tipo de documento: | Artigo |
Idioma: | eng |
Título da fonte: | Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos) |
Texto Completo: | http://hdl.handle.net/10316/100779 https://doi.org/10.3390/aerospace8040113 |
Resumo: | This paper presents a Reinforcement Learning (RL) approach to optimize the long-term scheduling of maintenance for an aircraft fleet. The problem considers fleet status, maintenance capacity, and other maintenance constraints to schedule hangar checks for a specified time horizon. The checks are scheduled within an interval, and the goal is to, schedule them as close as possible to their due date. In doing so, the number of checks is reduced, and the fleet availability increases. A Deep Q-learning algorithm is used to optimize the scheduling policy. The model is validated in a real scenario using maintenance data from 45 aircraft. The maintenance plan that is generated with our approach is compared with a previous study, which presented a Dynamic Programming (DP) based approach and airline estimations for the same period. The results show a reduction in the number of checks scheduled, which indicates the potential of RL in solving this problem. The adaptability of RL is also tested by introducing small disturbances in the initial conditions. After training the model with these simulated scenarios, the results show the robustness of the RL approach and its ability to generate efficient maintenance plans in only a few seconds. |
id |
RCAP_f0c77fe6a4a7d424fa06ab381778c712 |
---|---|
oai_identifier_str |
oai:estudogeral.uc.pt:10316/100779 |
network_acronym_str |
RCAP |
network_name_str |
Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos) |
repository_id_str |
7160 |
spelling |
Aircraft Maintenance Check Scheduling Using Reinforcement Learningaircraft maintenancemaintenance check schedulingreinforcement learningq-learningThis paper presents a Reinforcement Learning (RL) approach to optimize the long-term scheduling of maintenance for an aircraft fleet. The problem considers fleet status, maintenance capacity, and other maintenance constraints to schedule hangar checks for a specified time horizon. The checks are scheduled within an interval, and the goal is to, schedule them as close as possible to their due date. In doing so, the number of checks is reduced, and the fleet availability increases. A Deep Q-learning algorithm is used to optimize the scheduling policy. The model is validated in a real scenario using maintenance data from 45 aircraft. The maintenance plan that is generated with our approach is compared with a previous study, which presented a Dynamic Programming (DP) based approach and airline estimations for the same period. The results show a reduction in the number of checks scheduled, which indicates the potential of RL in solving this problem. The adaptability of RL is also tested by introducing small disturbances in the initial conditions. After training the model with these simulated scenarios, the results show the robustness of the RL approach and its ability to generate efficient maintenance plans in only a few seconds.2021info:eu-repo/semantics/publishedVersioninfo:eu-repo/semantics/articlehttp://hdl.handle.net/10316/100779http://hdl.handle.net/10316/100779https://doi.org/10.3390/aerospace8040113eng2226-4310Andrade, PedroSilva, CatarinaRibeiro, Bernardete MartinsSantos, Bruno F.info:eu-repo/semantics/openAccessreponame:Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos)instname:Agência para a Sociedade do Conhecimento (UMIC) - FCT - Sociedade da Informaçãoinstacron:RCAAP2022-11-29T15:58:06Zoai:estudogeral.uc.pt:10316/100779Portal AgregadorONGhttps://www.rcaap.pt/oai/openaireopendoar:71602024-03-19T21:18:05.634079Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos) - Agência para a Sociedade do Conhecimento (UMIC) - FCT - Sociedade da Informaçãofalse |
dc.title.none.fl_str_mv |
Aircraft Maintenance Check Scheduling Using Reinforcement Learning |
title |
Aircraft Maintenance Check Scheduling Using Reinforcement Learning |
spellingShingle |
Aircraft Maintenance Check Scheduling Using Reinforcement Learning Andrade, Pedro aircraft maintenance maintenance check scheduling reinforcement learning q-learning |
title_short |
Aircraft Maintenance Check Scheduling Using Reinforcement Learning |
title_full |
Aircraft Maintenance Check Scheduling Using Reinforcement Learning |
title_fullStr |
Aircraft Maintenance Check Scheduling Using Reinforcement Learning |
title_full_unstemmed |
Aircraft Maintenance Check Scheduling Using Reinforcement Learning |
title_sort |
Aircraft Maintenance Check Scheduling Using Reinforcement Learning |
author |
Andrade, Pedro |
author_facet |
Andrade, Pedro Silva, Catarina Ribeiro, Bernardete Martins Santos, Bruno F. |
author_role |
author |
author2 |
Silva, Catarina Ribeiro, Bernardete Martins Santos, Bruno F. |
author2_role |
author author author |
dc.contributor.author.fl_str_mv |
Andrade, Pedro Silva, Catarina Ribeiro, Bernardete Martins Santos, Bruno F. |
dc.subject.por.fl_str_mv |
aircraft maintenance maintenance check scheduling reinforcement learning q-learning |
topic |
aircraft maintenance maintenance check scheduling reinforcement learning q-learning |
description |
This paper presents a Reinforcement Learning (RL) approach to optimize the long-term scheduling of maintenance for an aircraft fleet. The problem considers fleet status, maintenance capacity, and other maintenance constraints to schedule hangar checks for a specified time horizon. The checks are scheduled within an interval, and the goal is to, schedule them as close as possible to their due date. In doing so, the number of checks is reduced, and the fleet availability increases. A Deep Q-learning algorithm is used to optimize the scheduling policy. The model is validated in a real scenario using maintenance data from 45 aircraft. The maintenance plan that is generated with our approach is compared with a previous study, which presented a Dynamic Programming (DP) based approach and airline estimations for the same period. The results show a reduction in the number of checks scheduled, which indicates the potential of RL in solving this problem. The adaptability of RL is also tested by introducing small disturbances in the initial conditions. After training the model with these simulated scenarios, the results show the robustness of the RL approach and its ability to generate efficient maintenance plans in only a few seconds. |
publishDate |
2021 |
dc.date.none.fl_str_mv |
2021 |
dc.type.status.fl_str_mv |
info:eu-repo/semantics/publishedVersion |
dc.type.driver.fl_str_mv |
info:eu-repo/semantics/article |
format |
article |
status_str |
publishedVersion |
dc.identifier.uri.fl_str_mv |
http://hdl.handle.net/10316/100779 http://hdl.handle.net/10316/100779 https://doi.org/10.3390/aerospace8040113 |
url |
http://hdl.handle.net/10316/100779 https://doi.org/10.3390/aerospace8040113 |
dc.language.iso.fl_str_mv |
eng |
language |
eng |
dc.relation.none.fl_str_mv |
2226-4310 |
dc.rights.driver.fl_str_mv |
info:eu-repo/semantics/openAccess |
eu_rights_str_mv |
openAccess |
dc.source.none.fl_str_mv |
reponame:Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos) instname:Agência para a Sociedade do Conhecimento (UMIC) - FCT - Sociedade da Informação instacron:RCAAP |
instname_str |
Agência para a Sociedade do Conhecimento (UMIC) - FCT - Sociedade da Informação |
instacron_str |
RCAAP |
institution |
RCAAP |
reponame_str |
Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos) |
collection |
Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos) |
repository.name.fl_str_mv |
Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos) - Agência para a Sociedade do Conhecimento (UMIC) - FCT - Sociedade da Informação |
repository.mail.fl_str_mv |
|
_version_ |
1799134076247998464 |