Síntese de fotografias e vídeos com depth-image-based rendering

Detalhes bibliográficos
Autor(a) principal: Oliveira, Adriano Quilião de
Data de Publicação: 2019
Tipo de documento: Tese
Idioma: por
Título da fonte: Biblioteca Digital de Teses e Dissertações da UFRGS
Texto Completo: http://hdl.handle.net/10183/201264
Resumo: The view synthesis process with Depth-Image-Based Rendering (DIBR) is presented as a promising way to enable applications like TV3D, Free Viewpoint Video, and others related to Virtual Reality and Augmented Reality. DIBR allows numerous virtual views of the same scene to be produced using only a single reference image and its depth map. However, artifacts (cracks and ghosts) and regions without information (holes) are formed in the synthesis process, which need to be treated or filled. In this thesis, we present two approaches for view synthesis with DIBR: ATA and DHS. We developed the ATA approach from in-depth studies on the generation of synthetic still images. This approach identifies empty and translucent cracks and reconstructs the affected regions with a specialized algorithm. Then, ghosts are identified through an evaluation process and warped to the correct positions. Finally, the remaining empty regions are filled with an adaptation of a popular inpainting algorithm that employs dynamically sized patches copied from the reference image and fits different hole types. The DHS approach, on the other hand, uses the advances produced with ATA, presenting an even more robust and reliable hole reconstruction method, aware of the image structure and composition. Ghosts are treated before the virtual view generation. An inpainting algorithm based on hierarchical superpixels is used for hole filling, which reconstructs empty regions based on their neighborhood composition by copying the contents of the reference image. Additionally, we propose a robust method for generating an incremental background model for videos that can be incorporated into any DIBR approach. As an example, we detailed its integration with DHS, which presented better results in the frame-by-frame evaluation. Exhaustive testing proves that the proposed approaches yield better quantitative and qualitative results when compared to several recent and competitive methods, both in the generation of synthetic still images as in videos, in tests with ground truth and real depth maps. As an additional contribution of this thesis, we evaluated the impact of using real depth maps produced with stereo matching in the synthesis process with DIBR and analyzed the relationship between quality metrics employed in both problems.
id URGS_4639e01358a24c8525df94865a78bfe4
oai_identifier_str oai:www.lume.ufrgs.br:10183/201264
network_acronym_str URGS
network_name_str Biblioteca Digital de Teses e Dissertações da UFRGS
repository_id_str 1853
spelling Oliveira, Adriano Quilião deWalter, MarceloJung, Claudio Rosito2019-11-02T03:51:40Z2019http://hdl.handle.net/10183/201264001103854The view synthesis process with Depth-Image-Based Rendering (DIBR) is presented as a promising way to enable applications like TV3D, Free Viewpoint Video, and others related to Virtual Reality and Augmented Reality. DIBR allows numerous virtual views of the same scene to be produced using only a single reference image and its depth map. However, artifacts (cracks and ghosts) and regions without information (holes) are formed in the synthesis process, which need to be treated or filled. In this thesis, we present two approaches for view synthesis with DIBR: ATA and DHS. We developed the ATA approach from in-depth studies on the generation of synthetic still images. This approach identifies empty and translucent cracks and reconstructs the affected regions with a specialized algorithm. Then, ghosts are identified through an evaluation process and warped to the correct positions. Finally, the remaining empty regions are filled with an adaptation of a popular inpainting algorithm that employs dynamically sized patches copied from the reference image and fits different hole types. The DHS approach, on the other hand, uses the advances produced with ATA, presenting an even more robust and reliable hole reconstruction method, aware of the image structure and composition. Ghosts are treated before the virtual view generation. An inpainting algorithm based on hierarchical superpixels is used for hole filling, which reconstructs empty regions based on their neighborhood composition by copying the contents of the reference image. Additionally, we propose a robust method for generating an incremental background model for videos that can be incorporated into any DIBR approach. As an example, we detailed its integration with DHS, which presented better results in the frame-by-frame evaluation. Exhaustive testing proves that the proposed approaches yield better quantitative and qualitative results when compared to several recent and competitive methods, both in the generation of synthetic still images as in videos, in tests with ground truth and real depth maps. As an additional contribution of this thesis, we evaluated the impact of using real depth maps produced with stereo matching in the synthesis process with DIBR and analyzed the relationship between quality metrics employed in both problems.O processo de síntese de vistas com Depth-Image-Based Rendering (DIBR) se apresenta como um meio promissor para viabilizar aplicações como TV3D, Free Viewpoint Video, e outras relacionadas com Realidade Virtual e Realidade Aumentada. DIBR permite que sejam produzidos inúmeros pontos de vista virtuais da mesma cena utilizando apenas uma imagem de referência e seu respectivo mapa de profundidades. Contudo, artefatos (cracks e ghosts) e regiões sem informação (holes) são formados no processo de síntese, os quais precisam ser tratados ou preenchidos. Nesta tese, são apresentadas duas abordagens para a síntese de vistas com DIBR: ATA e DHS. A abordagem ATA foi desenvolvida a partir de estudos aprofundados acerca da geração de fotografias sintéticas. Esta identifica cracks vazios e translúcidos e reconstrói as regiões afetadas com um algoritmo especializado. Ghosts são identificados e reprojetados para as posições corretas de acordo com um processo de avaliação. As regiões sem informação restantes são preenchidas com uma adaptação de um popular algoritmo de inpainting, que emprega patches com tamanho dinâmico, copiados da imagem de referência, e que se ajusta a diferentes tipos de hole. Já a abordagem DHS utiliza os avanços produzidos com ATA, apresentando um método de reconstrução dos holes ainda mais robusto e confiável, ciente da estrutura e composição da imagem. Ghosts são tratados antes da geração da vista virtual. Para os holes, utilizase um algoritmo de inpainting baseado em superpixels hierárquicos, que reconstrói as regiões vazias com base na composição de sua vizinhança, copiando conteúdo da imagem de referência. Adicionalmente, propõe-se um método robusto para a geração de um modelo de background incremental para vídeos, que pode ser incorporado em qualquer abordagem DIBR. Como exemplo, detalha-se sua integração ao DHS, que apresentou melhores resultados na avaliação quadro a quadro realizada. Exaustivos testes comprovam que as abordagens propostas apresentam melhores resultados quantitativos e qualitativos quando comparados com diversos métodos recentes e competitivos, tanto na geração de fotografias como de vídeos sintéticos, em testes com mapas de profundidade ground truth e reais. Como contribuição adicional desta tese, avaliou-se o impacto do uso de mapas de profundidade reais produzidos com casamento estéreo no processo de síntese com DIBR, e analisou-se a relação entre métricas de qualidade empregadas em ambos problemas.application/pdfporComputação gráficaProcessamento de imagensVídeo digitalview synthesisinpaintingtemporal informationSíntese de fotografias e vídeos com depth-image-based renderingSynthesis of still images and videos with depth-image-based rendering info:eu-repo/semantics/publishedVersioninfo:eu-repo/semantics/doctoralThesisUniversidade Federal do Rio Grande do SulInstituto de InformáticaPrograma de Pós-Graduação em ComputaçãoPorto Alegre, BR-RS2019doutoradoinfo:eu-repo/semantics/openAccessreponame:Biblioteca Digital de Teses e Dissertações da UFRGSinstname:Universidade Federal do Rio Grande do Sul (UFRGS)instacron:UFRGSTEXT001103854.pdf.txt001103854.pdf.txtExtracted Texttext/plain325193http://www.lume.ufrgs.br/bitstream/10183/201264/2/001103854.pdf.txt82a44d0b5b80b88afd7c4e59bb2f1481MD52ORIGINAL001103854.pdfTexto completoapplication/pdf2528160http://www.lume.ufrgs.br/bitstream/10183/201264/1/001103854.pdfff3e733dc94be1d9314be6d20c7ad2c4MD5110183/2012642021-05-26 04:35:09.992099oai:www.lume.ufrgs.br:10183/201264Biblioteca Digital de Teses e Dissertaçõeshttps://lume.ufrgs.br/handle/10183/2PUBhttps://lume.ufrgs.br/oai/requestlume@ufrgs.br||lume@ufrgs.bropendoar:18532021-05-26T07:35:09Biblioteca Digital de Teses e Dissertações da UFRGS - Universidade Federal do Rio Grande do Sul (UFRGS)false
dc.title.pt_BR.fl_str_mv Síntese de fotografias e vídeos com depth-image-based rendering
dc.title.alternative.en.fl_str_mv Synthesis of still images and videos with depth-image-based rendering
title Síntese de fotografias e vídeos com depth-image-based rendering
spellingShingle Síntese de fotografias e vídeos com depth-image-based rendering
Oliveira, Adriano Quilião de
Computação gráfica
Processamento de imagens
Vídeo digital
view synthesis
inpainting
temporal information
title_short Síntese de fotografias e vídeos com depth-image-based rendering
title_full Síntese de fotografias e vídeos com depth-image-based rendering
title_fullStr Síntese de fotografias e vídeos com depth-image-based rendering
title_full_unstemmed Síntese de fotografias e vídeos com depth-image-based rendering
title_sort Síntese de fotografias e vídeos com depth-image-based rendering
author Oliveira, Adriano Quilião de
author_facet Oliveira, Adriano Quilião de
author_role author
dc.contributor.author.fl_str_mv Oliveira, Adriano Quilião de
dc.contributor.advisor1.fl_str_mv Walter, Marcelo
dc.contributor.advisor-co1.fl_str_mv Jung, Claudio Rosito
contributor_str_mv Walter, Marcelo
Jung, Claudio Rosito
dc.subject.por.fl_str_mv Computação gráfica
Processamento de imagens
Vídeo digital
topic Computação gráfica
Processamento de imagens
Vídeo digital
view synthesis
inpainting
temporal information
dc.subject.eng.fl_str_mv view synthesis
inpainting
temporal information
description The view synthesis process with Depth-Image-Based Rendering (DIBR) is presented as a promising way to enable applications like TV3D, Free Viewpoint Video, and others related to Virtual Reality and Augmented Reality. DIBR allows numerous virtual views of the same scene to be produced using only a single reference image and its depth map. However, artifacts (cracks and ghosts) and regions without information (holes) are formed in the synthesis process, which need to be treated or filled. In this thesis, we present two approaches for view synthesis with DIBR: ATA and DHS. We developed the ATA approach from in-depth studies on the generation of synthetic still images. This approach identifies empty and translucent cracks and reconstructs the affected regions with a specialized algorithm. Then, ghosts are identified through an evaluation process and warped to the correct positions. Finally, the remaining empty regions are filled with an adaptation of a popular inpainting algorithm that employs dynamically sized patches copied from the reference image and fits different hole types. The DHS approach, on the other hand, uses the advances produced with ATA, presenting an even more robust and reliable hole reconstruction method, aware of the image structure and composition. Ghosts are treated before the virtual view generation. An inpainting algorithm based on hierarchical superpixels is used for hole filling, which reconstructs empty regions based on their neighborhood composition by copying the contents of the reference image. Additionally, we propose a robust method for generating an incremental background model for videos that can be incorporated into any DIBR approach. As an example, we detailed its integration with DHS, which presented better results in the frame-by-frame evaluation. Exhaustive testing proves that the proposed approaches yield better quantitative and qualitative results when compared to several recent and competitive methods, both in the generation of synthetic still images as in videos, in tests with ground truth and real depth maps. As an additional contribution of this thesis, we evaluated the impact of using real depth maps produced with stereo matching in the synthesis process with DIBR and analyzed the relationship between quality metrics employed in both problems.
publishDate 2019
dc.date.accessioned.fl_str_mv 2019-11-02T03:51:40Z
dc.date.issued.fl_str_mv 2019
dc.type.status.fl_str_mv info:eu-repo/semantics/publishedVersion
dc.type.driver.fl_str_mv info:eu-repo/semantics/doctoralThesis
format doctoralThesis
status_str publishedVersion
dc.identifier.uri.fl_str_mv http://hdl.handle.net/10183/201264
dc.identifier.nrb.pt_BR.fl_str_mv 001103854
url http://hdl.handle.net/10183/201264
identifier_str_mv 001103854
dc.language.iso.fl_str_mv por
language por
dc.rights.driver.fl_str_mv info:eu-repo/semantics/openAccess
eu_rights_str_mv openAccess
dc.format.none.fl_str_mv application/pdf
dc.source.none.fl_str_mv reponame:Biblioteca Digital de Teses e Dissertações da UFRGS
instname:Universidade Federal do Rio Grande do Sul (UFRGS)
instacron:UFRGS
instname_str Universidade Federal do Rio Grande do Sul (UFRGS)
instacron_str UFRGS
institution UFRGS
reponame_str Biblioteca Digital de Teses e Dissertações da UFRGS
collection Biblioteca Digital de Teses e Dissertações da UFRGS
bitstream.url.fl_str_mv http://www.lume.ufrgs.br/bitstream/10183/201264/2/001103854.pdf.txt
http://www.lume.ufrgs.br/bitstream/10183/201264/1/001103854.pdf
bitstream.checksum.fl_str_mv 82a44d0b5b80b88afd7c4e59bb2f1481
ff3e733dc94be1d9314be6d20c7ad2c4
bitstream.checksumAlgorithm.fl_str_mv MD5
MD5
repository.name.fl_str_mv Biblioteca Digital de Teses e Dissertações da UFRGS - Universidade Federal do Rio Grande do Sul (UFRGS)
repository.mail.fl_str_mv lume@ufrgs.br||lume@ufrgs.br
_version_ 1800309155747069952