An instrument for reviewing the completeness of experimental plans for controlled experiments using human subjects in software engineering

Detalhes bibliográficos
Autor(a) principal: FONSECA, Liliane Sheyla da Silva
Data de Publicação: 2016
Tipo de documento: Tese
Idioma: eng
Título da fonte: Repositório Institucional da UFPE
dARK ID: ark:/64986/0013000010mf1
Texto Completo: https://repositorio.ufpe.br/handle/123456789/22421
Resumo: It is common sense in software engineering that well made experimental plans are recipes for successful experiments, and they help experimenters to avoid interferences during experiments. Although a number of tools are available to help researchers with writing experiments reports for scientific publications, few studies focus on how to assess study protocols with respect to completeness and scientific quality. As a result, designing controlled experiments using subjects has been a challenge for many experimenters in software engineering because of a large variety of factors that should be present in it to avoid introducing bias in controlled experiments. The main aim of this thesis is to define an instrument to help experimenters, specially beginners, to review their experimental planning for assessing whether they produced an experimental plan that is complete and includes all possible factors to minimize bias and issues. The instrument is a checklist whose design is based on experimental best practices and experts’ experience in planning and conducting controlled experiments using subjects. To collect the best practices, a systematic mapping study was conducted to identify support mechanisms (processes, tools, guidelines, among others.) used to plan and conduct empirical studies in the empirical software engineering community, and an informal literature review was carried out in order to find which support mechanisms are generally used in other fields. Moreover, we performed a qualitative study for understanding how empirical software engineering experts plan their experiments. The instrument has been evaluated through four empirical studies. Each one was explored from different perspectives by Software Engineering researchers at different levels of experience. The instrument was assessed regarding items that they find useful, inter-rater agreement, inter-rater reliability, and criterion validity using fully crossed design. Two controlled experiments were performed to assess if the usage of the instrument can reduce the chance of forgetting to include something important during the experiment planning phase compared to the usage of ad hoc practices. Additionally, the acceptance of the instrument was assessed by the four studies. In total, we had 35 participants who participated in four different kinds of assesment of the instrument. In the first study, 75.76% of the items were judged useful by two experts. The remaining items were discussed and adjusted. The second study revealed that the usage of the instrument helped beginners to assess experimental plans in the same way as the experts. We found a strong correlation between the overall completeness scores of the experimental plans and the recommendation that the experiment should proceed or not, and whether it is likely to be successful. In Studies 3 and 4, the proportion of the correct items found by participants using the instrument was greater than the results from participants using the ad hoc practices. The instrument has high acceptance from participants. Although the results are positive, performing more assessments including different settings is required to generalize these results. The usage of the instrument by experimenters, specially beginners, helps them to review the key factors included in the experimental plan, thus contributing to reduce potential confounding factors in the experiment. Revising an experimental plan is not a direct evaluation of the quality of the experiment itself but it allows changes to be made to improve the experiment before it is performed.
id UFPE_7d4c340bd36a9add72e4fb3da7c889d9
oai_identifier_str oai:repositorio.ufpe.br:123456789/22421
network_acronym_str UFPE
network_name_str Repositório Institucional da UFPE
repository_id_str 2221
spelling FONSECA, Liliane Sheyla da Silvahttp://lattes.cnpq.br/9581301411261992http://lattes.cnpq.br/6456667887502521SOARES, Sergio Castelo BrancoSEAMAN, Carolyn2017-11-29T17:23:55Z2017-11-29T17:23:55Z2016-12-21https://repositorio.ufpe.br/handle/123456789/22421ark:/64986/0013000010mf1It is common sense in software engineering that well made experimental plans are recipes for successful experiments, and they help experimenters to avoid interferences during experiments. Although a number of tools are available to help researchers with writing experiments reports for scientific publications, few studies focus on how to assess study protocols with respect to completeness and scientific quality. As a result, designing controlled experiments using subjects has been a challenge for many experimenters in software engineering because of a large variety of factors that should be present in it to avoid introducing bias in controlled experiments. The main aim of this thesis is to define an instrument to help experimenters, specially beginners, to review their experimental planning for assessing whether they produced an experimental plan that is complete and includes all possible factors to minimize bias and issues. The instrument is a checklist whose design is based on experimental best practices and experts’ experience in planning and conducting controlled experiments using subjects. To collect the best practices, a systematic mapping study was conducted to identify support mechanisms (processes, tools, guidelines, among others.) used to plan and conduct empirical studies in the empirical software engineering community, and an informal literature review was carried out in order to find which support mechanisms are generally used in other fields. Moreover, we performed a qualitative study for understanding how empirical software engineering experts plan their experiments. The instrument has been evaluated through four empirical studies. Each one was explored from different perspectives by Software Engineering researchers at different levels of experience. The instrument was assessed regarding items that they find useful, inter-rater agreement, inter-rater reliability, and criterion validity using fully crossed design. Two controlled experiments were performed to assess if the usage of the instrument can reduce the chance of forgetting to include something important during the experiment planning phase compared to the usage of ad hoc practices. Additionally, the acceptance of the instrument was assessed by the four studies. In total, we had 35 participants who participated in four different kinds of assesment of the instrument. In the first study, 75.76% of the items were judged useful by two experts. The remaining items were discussed and adjusted. The second study revealed that the usage of the instrument helped beginners to assess experimental plans in the same way as the experts. We found a strong correlation between the overall completeness scores of the experimental plans and the recommendation that the experiment should proceed or not, and whether it is likely to be successful. In Studies 3 and 4, the proportion of the correct items found by participants using the instrument was greater than the results from participants using the ad hoc practices. The instrument has high acceptance from participants. Although the results are positive, performing more assessments including different settings is required to generalize these results. The usage of the instrument by experimenters, specially beginners, helps them to review the key factors included in the experimental plan, thus contributing to reduce potential confounding factors in the experiment. Revising an experimental plan is not a direct evaluation of the quality of the experiment itself but it allows changes to be made to improve the experiment before it is performed.FACEPEÉ comumente aceito pela comunidade de engenharia de software que planos experimentais bem planejados são receitas para experimentos bem sucedidos. Isso se deve ao fato que planos experimentais auxiliam experimentadores a evitarem interferências durante a execução dos experimentos. No entanto, embora existam ferramentas disponíveis para ajudar os investigadores a reportarem seus experimentos para publicações científicas, poucos estudos tem o objetivo de avaliar os protocolos de estudo no que diz respeito à completude e qualidade científica. Desta forma, planejar experimentos controlados utilizando participantes tem sido um desafio para muitos experimentadores em engenharia de software devido a grande variedade de fatores que devem estar presentes em um plano experimental a fim de evitar a introdução de viés nos experimentos controlados. O principal objetivo dessa tese de doutorado é definir um instrumento que auxilie experimentadores, principalmente inexperientes, a revisarem seus planejamentos experimentais a fim de avaliar se eles produziram um plano experimental completo, incluindo todos os possíveis fatores para minimizar viés e problemas. O instrumento é uma lista de verificação baseado nas melhores práticas experimentais e na experiência dos especialistas em engenharia de software experimental no planejamento e condução de experimentos controlados utilizando pessoas. Para coletar as melhores práticas, um mapeamento sistemático foi realizado para identificar os mecanismos de apoio (processos, ferramentas, guias, dentre outros) utilizados para planejar e conduzir estudos empíricos na comunidade de engenharia de software e uma revisão da literatura foi realizada para identificar mecanismos de apoio que são geralmente utilizados em outras áreas. Além disso, foi realizada um estudo qualitativo a fim de entender como os especialistas em engenharia de software experimental planejam seus experimentos. O instrumento foi avaliado por meio de quatro estudos. Cada estudo foi explorado através de diferentes perspectivas por pesquisadores de engenharia de software em diferentes níveis de experiência. O instrumento foi avaliado com relação a utilidade dos itens, a concordância e a confiabilidade entre os avaliadores e validade de critério. Dois experimentos controlados foram realizados para avaliar se o uso do instrumento pode reduzir a chance de esquecer algo importante durante a fase de planejamento do experimento em comparação com as práticas comumente usadas pelos pesquisadores. Além disso, os quatro estudos avaliaram a aceitação do instrumento para revisar planos experimentais de experimentos controlados utilizando participantes. No total, 35 participantes avaliaram o instrumento através de quatro diferentes tipos de objetivos. No primeiro estudo, 75,76% dos itens foram julgados uteis pelos dois especialistas envolvidos no estudo. Os itens restantes foram discutidos e ajustados. O segundo estudo revelou que a utilização do instrumento auxiliou iniciantes a avaliarem planos experimentais da mesma forma dos especialistas. Os resultados mostraram uma forte correlação entre os escores da completude global dos planos experimentais e as recomendações se o experimento deveria prosseguir e a probabilidade do experimento ser bem-sucedido. Nos estudos 3 e 4, a proporção dos itens corretos encontrados pelos participantes utilizando o instrumento foi significantemente maior do que os resultados utilizando as práticas comumente utilizadas pelos participantes. O instrumento teve alta aceitação por parte dos participantes. No entanto, embora os resultados sejam positivos, se faz necessário a realização de mais estudos de avaliação, incluindo outras configurações de ambientes a fim de que o resultado possa ser generalizado. A utilização do instrumento pelos experimentadores, especialmente os iniciantes, auxilia a revisão dos principais fatores que devem estar incluídos no plano experimental, contribuindo assim para reduzir potenciais fatores de confusão no experimento. Revisar um plano experimental não é uma avaliação direta da qualidade do experimento, mas permite que mudanças no experimento sejam realizadas antes que ele seja de fato executado.engUniversidade Federal de PernambucoPrograma de Pos Graduacao em Ciencia da ComputacaoUFPEBrasilAttribution-NonCommercial-NoDerivs 3.0 Brazilhttp://creativecommons.org/licenses/by-nc-nd/3.0/br/info:eu-repo/semantics/openAccessEngenharia de softwareEngenharia de software experimentalExperimentos controladosFatores humanosAn instrument for reviewing the completeness of experimental plans for controlled experiments using human subjects in software engineeringinfo:eu-repo/semantics/publishedVersioninfo:eu-repo/semantics/doctoralThesisdoutoradoreponame:Repositório Institucional da UFPEinstname:Universidade Federal de Pernambuco (UFPE)instacron:UFPETHUMBNAILTeseDoutorado_lssf.pdf.jpgTeseDoutorado_lssf.pdf.jpgGenerated Thumbnailimage/jpeg1277https://repositorio.ufpe.br/bitstream/123456789/22421/5/TeseDoutorado_lssf.pdf.jpg9e6ca26e86e1da18b02cf378d7190192MD55ORIGINALTeseDoutorado_lssf.pdfTeseDoutorado_lssf.pdfapplication/pdf10022088https://repositorio.ufpe.br/bitstream/123456789/22421/1/TeseDoutorado_lssf.pdfda65676610d3186199f2a0ed06e7f8f1MD51CC-LICENSElicense_rdflicense_rdfapplication/rdf+xml; charset=utf-8811https://repositorio.ufpe.br/bitstream/123456789/22421/2/license_rdfe39d27027a6cc9cb039ad269a5db8e34MD52LICENSElicense.txtlicense.txttext/plain; charset=utf-82311https://repositorio.ufpe.br/bitstream/123456789/22421/3/license.txt4b8a02c7f2818eaf00dcf2260dd5eb08MD53TEXTTeseDoutorado_lssf.pdf.txtTeseDoutorado_lssf.pdf.txtExtracted texttext/plain575104https://repositorio.ufpe.br/bitstream/123456789/22421/4/TeseDoutorado_lssf.pdf.txt88376c5a03d64f5e1d83294be238940fMD54123456789/224212019-10-25 08:03:32.506oai:repositorio.ufpe.br:123456789/22421TGljZW7Dp2EgZGUgRGlzdHJpYnVpw6fDo28gTsOjbyBFeGNsdXNpdmEKClRvZG8gZGVwb3NpdGFudGUgZGUgbWF0ZXJpYWwgbm8gUmVwb3NpdMOzcmlvIEluc3RpdHVjaW9uYWwgKFJJKSBkZXZlIGNvbmNlZGVyLCDDoCBVbml2ZXJzaWRhZGUgRmVkZXJhbCBkZSBQZXJuYW1idWNvIChVRlBFKSwgdW1hIExpY2Vuw6dhIGRlIERpc3RyaWJ1acOnw6NvIE7Do28gRXhjbHVzaXZhIHBhcmEgbWFudGVyIGUgdG9ybmFyIGFjZXNzw612ZWlzIG9zIHNldXMgZG9jdW1lbnRvcywgZW0gZm9ybWF0byBkaWdpdGFsLCBuZXN0ZSByZXBvc2l0w7NyaW8uCgpDb20gYSBjb25jZXNzw6NvIGRlc3RhIGxpY2Vuw6dhIG7Do28gZXhjbHVzaXZhLCBvIGRlcG9zaXRhbnRlIG1hbnTDqW0gdG9kb3Mgb3MgZGlyZWl0b3MgZGUgYXV0b3IuCl9fX19fX19fX19fX19fX19fX19fX19fX19fX19fX19fX19fX19fX19fX19fX19fX19fX19fX19fX19fX19fX19fX19fXwoKTGljZW7Dp2EgZGUgRGlzdHJpYnVpw6fDo28gTsOjbyBFeGNsdXNpdmEKCkFvIGNvbmNvcmRhciBjb20gZXN0YSBsaWNlbsOnYSBlIGFjZWl0w6EtbGEsIHZvY8OqIChhdXRvciBvdSBkZXRlbnRvciBkb3MgZGlyZWl0b3MgYXV0b3JhaXMpOgoKYSkgRGVjbGFyYSBxdWUgY29uaGVjZSBhIHBvbMOtdGljYSBkZSBjb3B5cmlnaHQgZGEgZWRpdG9yYSBkbyBzZXUgZG9jdW1lbnRvOwpiKSBEZWNsYXJhIHF1ZSBjb25oZWNlIGUgYWNlaXRhIGFzIERpcmV0cml6ZXMgcGFyYSBvIFJlcG9zaXTDs3JpbyBJbnN0aXR1Y2lvbmFsIGRhIFVGUEU7CmMpIENvbmNlZGUgw6AgVUZQRSBvIGRpcmVpdG8gbsOjbyBleGNsdXNpdm8gZGUgYXJxdWl2YXIsIHJlcHJvZHV6aXIsIGNvbnZlcnRlciAoY29tbyBkZWZpbmlkbyBhIHNlZ3VpciksIGNvbXVuaWNhciBlL291IGRpc3RyaWJ1aXIsIG5vIFJJLCBvIGRvY3VtZW50byBlbnRyZWd1ZSAoaW5jbHVpbmRvIG8gcmVzdW1vL2Fic3RyYWN0KSBlbSBmb3JtYXRvIGRpZ2l0YWwgb3UgcG9yIG91dHJvIG1laW87CmQpIERlY2xhcmEgcXVlIGF1dG9yaXphIGEgVUZQRSBhIGFycXVpdmFyIG1haXMgZGUgdW1hIGPDs3BpYSBkZXN0ZSBkb2N1bWVudG8gZSBjb252ZXJ0w6otbG8sIHNlbSBhbHRlcmFyIG8gc2V1IGNvbnRlw7pkbywgcGFyYSBxdWFscXVlciBmb3JtYXRvIGRlIGZpY2hlaXJvLCBtZWlvIG91IHN1cG9ydGUsIHBhcmEgZWZlaXRvcyBkZSBzZWd1cmFuw6dhLCBwcmVzZXJ2YcOnw6NvIChiYWNrdXApIGUgYWNlc3NvOwplKSBEZWNsYXJhIHF1ZSBvIGRvY3VtZW50byBzdWJtZXRpZG8gw6kgbyBzZXUgdHJhYmFsaG8gb3JpZ2luYWwgZSBxdWUgZGV0w6ltIG8gZGlyZWl0byBkZSBjb25jZWRlciBhIHRlcmNlaXJvcyBvcyBkaXJlaXRvcyBjb250aWRvcyBuZXN0YSBsaWNlbsOnYS4gRGVjbGFyYSB0YW1iw6ltIHF1ZSBhIGVudHJlZ2EgZG8gZG9jdW1lbnRvIG7Do28gaW5mcmluZ2Ugb3MgZGlyZWl0b3MgZGUgb3V0cmEgcGVzc29hIG91IGVudGlkYWRlOwpmKSBEZWNsYXJhIHF1ZSwgbm8gY2FzbyBkbyBkb2N1bWVudG8gc3VibWV0aWRvIGNvbnRlciBtYXRlcmlhbCBkbyBxdWFsIG7Do28gZGV0w6ltIG9zIGRpcmVpdG9zIGRlCmF1dG9yLCBvYnRldmUgYSBhdXRvcml6YcOnw6NvIGlycmVzdHJpdGEgZG8gcmVzcGVjdGl2byBkZXRlbnRvciBkZXNzZXMgZGlyZWl0b3MgcGFyYSBjZWRlciDDoApVRlBFIG9zIGRpcmVpdG9zIHJlcXVlcmlkb3MgcG9yIGVzdGEgTGljZW7Dp2EgZSBhdXRvcml6YXIgYSB1bml2ZXJzaWRhZGUgYSB1dGlsaXrDoS1sb3MgbGVnYWxtZW50ZS4gRGVjbGFyYSB0YW1iw6ltIHF1ZSBlc3NlIG1hdGVyaWFsIGN1am9zIGRpcmVpdG9zIHPDo28gZGUgdGVyY2Vpcm9zIGVzdMOhIGNsYXJhbWVudGUgaWRlbnRpZmljYWRvIGUgcmVjb25oZWNpZG8gbm8gdGV4dG8gb3UgY29udGXDumRvIGRvIGRvY3VtZW50byBlbnRyZWd1ZTsKZykgU2UgbyBkb2N1bWVudG8gZW50cmVndWUgw6kgYmFzZWFkbyBlbSB0cmFiYWxobyBmaW5hbmNpYWRvIG91IGFwb2lhZG8gcG9yIG91dHJhIGluc3RpdHVpw6fDo28gcXVlIG7Do28gYSBVRlBFLMKgZGVjbGFyYSBxdWUgY3VtcHJpdSBxdWFpc3F1ZXIgb2JyaWdhw6fDtWVzIGV4aWdpZGFzIHBlbG8gcmVzcGVjdGl2byBjb250cmF0byBvdSBhY29yZG8uCgpBIFVGUEUgaWRlbnRpZmljYXLDoSBjbGFyYW1lbnRlIG8ocykgbm9tZShzKSBkbyhzKSBhdXRvciAoZXMpIGRvcyBkaXJlaXRvcyBkbyBkb2N1bWVudG8gZW50cmVndWUgZSBuw6NvIGZhcsOhIHF1YWxxdWVyIGFsdGVyYcOnw6NvLCBwYXJhIGFsw6ltIGRvIHByZXZpc3RvIG5hIGFsw61uZWEgYykuCg==Repositório InstitucionalPUBhttps://repositorio.ufpe.br/oai/requestattena@ufpe.bropendoar:22212019-10-25T11:03:32Repositório Institucional da UFPE - Universidade Federal de Pernambuco (UFPE)false
dc.title.pt_BR.fl_str_mv An instrument for reviewing the completeness of experimental plans for controlled experiments using human subjects in software engineering
title An instrument for reviewing the completeness of experimental plans for controlled experiments using human subjects in software engineering
spellingShingle An instrument for reviewing the completeness of experimental plans for controlled experiments using human subjects in software engineering
FONSECA, Liliane Sheyla da Silva
Engenharia de software
Engenharia de software experimental
Experimentos controlados
Fatores humanos
title_short An instrument for reviewing the completeness of experimental plans for controlled experiments using human subjects in software engineering
title_full An instrument for reviewing the completeness of experimental plans for controlled experiments using human subjects in software engineering
title_fullStr An instrument for reviewing the completeness of experimental plans for controlled experiments using human subjects in software engineering
title_full_unstemmed An instrument for reviewing the completeness of experimental plans for controlled experiments using human subjects in software engineering
title_sort An instrument for reviewing the completeness of experimental plans for controlled experiments using human subjects in software engineering
author FONSECA, Liliane Sheyla da Silva
author_facet FONSECA, Liliane Sheyla da Silva
author_role author
dc.contributor.authorLattes.pt_BR.fl_str_mv http://lattes.cnpq.br/9581301411261992
dc.contributor.advisorLattes.pt_BR.fl_str_mv http://lattes.cnpq.br/6456667887502521
dc.contributor.author.fl_str_mv FONSECA, Liliane Sheyla da Silva
dc.contributor.advisor1.fl_str_mv SOARES, Sergio Castelo Branco
dc.contributor.advisor-co1.fl_str_mv SEAMAN, Carolyn
contributor_str_mv SOARES, Sergio Castelo Branco
SEAMAN, Carolyn
dc.subject.por.fl_str_mv Engenharia de software
Engenharia de software experimental
Experimentos controlados
Fatores humanos
topic Engenharia de software
Engenharia de software experimental
Experimentos controlados
Fatores humanos
description It is common sense in software engineering that well made experimental plans are recipes for successful experiments, and they help experimenters to avoid interferences during experiments. Although a number of tools are available to help researchers with writing experiments reports for scientific publications, few studies focus on how to assess study protocols with respect to completeness and scientific quality. As a result, designing controlled experiments using subjects has been a challenge for many experimenters in software engineering because of a large variety of factors that should be present in it to avoid introducing bias in controlled experiments. The main aim of this thesis is to define an instrument to help experimenters, specially beginners, to review their experimental planning for assessing whether they produced an experimental plan that is complete and includes all possible factors to minimize bias and issues. The instrument is a checklist whose design is based on experimental best practices and experts’ experience in planning and conducting controlled experiments using subjects. To collect the best practices, a systematic mapping study was conducted to identify support mechanisms (processes, tools, guidelines, among others.) used to plan and conduct empirical studies in the empirical software engineering community, and an informal literature review was carried out in order to find which support mechanisms are generally used in other fields. Moreover, we performed a qualitative study for understanding how empirical software engineering experts plan their experiments. The instrument has been evaluated through four empirical studies. Each one was explored from different perspectives by Software Engineering researchers at different levels of experience. The instrument was assessed regarding items that they find useful, inter-rater agreement, inter-rater reliability, and criterion validity using fully crossed design. Two controlled experiments were performed to assess if the usage of the instrument can reduce the chance of forgetting to include something important during the experiment planning phase compared to the usage of ad hoc practices. Additionally, the acceptance of the instrument was assessed by the four studies. In total, we had 35 participants who participated in four different kinds of assesment of the instrument. In the first study, 75.76% of the items were judged useful by two experts. The remaining items were discussed and adjusted. The second study revealed that the usage of the instrument helped beginners to assess experimental plans in the same way as the experts. We found a strong correlation between the overall completeness scores of the experimental plans and the recommendation that the experiment should proceed or not, and whether it is likely to be successful. In Studies 3 and 4, the proportion of the correct items found by participants using the instrument was greater than the results from participants using the ad hoc practices. The instrument has high acceptance from participants. Although the results are positive, performing more assessments including different settings is required to generalize these results. The usage of the instrument by experimenters, specially beginners, helps them to review the key factors included in the experimental plan, thus contributing to reduce potential confounding factors in the experiment. Revising an experimental plan is not a direct evaluation of the quality of the experiment itself but it allows changes to be made to improve the experiment before it is performed.
publishDate 2016
dc.date.issued.fl_str_mv 2016-12-21
dc.date.accessioned.fl_str_mv 2017-11-29T17:23:55Z
dc.date.available.fl_str_mv 2017-11-29T17:23:55Z
dc.type.status.fl_str_mv info:eu-repo/semantics/publishedVersion
dc.type.driver.fl_str_mv info:eu-repo/semantics/doctoralThesis
format doctoralThesis
status_str publishedVersion
dc.identifier.uri.fl_str_mv https://repositorio.ufpe.br/handle/123456789/22421
dc.identifier.dark.fl_str_mv ark:/64986/0013000010mf1
url https://repositorio.ufpe.br/handle/123456789/22421
identifier_str_mv ark:/64986/0013000010mf1
dc.language.iso.fl_str_mv eng
language eng
dc.rights.driver.fl_str_mv Attribution-NonCommercial-NoDerivs 3.0 Brazil
http://creativecommons.org/licenses/by-nc-nd/3.0/br/
info:eu-repo/semantics/openAccess
rights_invalid_str_mv Attribution-NonCommercial-NoDerivs 3.0 Brazil
http://creativecommons.org/licenses/by-nc-nd/3.0/br/
eu_rights_str_mv openAccess
dc.publisher.none.fl_str_mv Universidade Federal de Pernambuco
dc.publisher.program.fl_str_mv Programa de Pos Graduacao em Ciencia da Computacao
dc.publisher.initials.fl_str_mv UFPE
dc.publisher.country.fl_str_mv Brasil
publisher.none.fl_str_mv Universidade Federal de Pernambuco
dc.source.none.fl_str_mv reponame:Repositório Institucional da UFPE
instname:Universidade Federal de Pernambuco (UFPE)
instacron:UFPE
instname_str Universidade Federal de Pernambuco (UFPE)
instacron_str UFPE
institution UFPE
reponame_str Repositório Institucional da UFPE
collection Repositório Institucional da UFPE
bitstream.url.fl_str_mv https://repositorio.ufpe.br/bitstream/123456789/22421/5/TeseDoutorado_lssf.pdf.jpg
https://repositorio.ufpe.br/bitstream/123456789/22421/1/TeseDoutorado_lssf.pdf
https://repositorio.ufpe.br/bitstream/123456789/22421/2/license_rdf
https://repositorio.ufpe.br/bitstream/123456789/22421/3/license.txt
https://repositorio.ufpe.br/bitstream/123456789/22421/4/TeseDoutorado_lssf.pdf.txt
bitstream.checksum.fl_str_mv 9e6ca26e86e1da18b02cf378d7190192
da65676610d3186199f2a0ed06e7f8f1
e39d27027a6cc9cb039ad269a5db8e34
4b8a02c7f2818eaf00dcf2260dd5eb08
88376c5a03d64f5e1d83294be238940f
bitstream.checksumAlgorithm.fl_str_mv MD5
MD5
MD5
MD5
MD5
repository.name.fl_str_mv Repositório Institucional da UFPE - Universidade Federal de Pernambuco (UFPE)
repository.mail.fl_str_mv attena@ufpe.br
_version_ 1815172964410195968