ERG-ARCH : a reinforcement learning architecture for propositionally constrained multi-agent state spaces

Bibliographic Details
Main Author: Anderson Viçoso de Araújo
Publication Date: 2014
Format: Doctoral thesis
Language: eng
Source: Biblioteca Digital de Teses e Dissertações do ITA
Download full: http://www.bd.bibl.ita.br/tde_busca/arquivo.php?codArquivo=3096
Summary: The main goal of this work is to present an approach that ?nds an appropriate set of sequential actions for a group of cooperative agents interacting over a constrained environment. This search is considered a complex task for autonomous agents and is not possible to use default reinforcement learning algorithms to learn the adequate policy. In this thesis, a technique that deals with propositionally constrained state spaces and makes use of a Reinforcement Learning algorithm based on Markov Decision Process is proposed. A new model is also presented which formally de?nes this restricted search space. By so doing, this work aims at reducing the overall exploratory need, thus improving the performance of the learning algorithm. To constrain the state space the concept of extended reachability goals is employed. Through them it is possible to de?ne an objective to be preserved during the iteration with the environment and another that de?nes a goal state. In this cooperative environment, the information about the propositions is shared among the agents during its interaction. An architecture to solve problems in such environments is also presented. Experiments to validate the proposed algorithm were performed on different test cases and showed interesting results. A performance evaluation against standard Reinforcement Learning techniques showed that by extending autonomous learning with propositional constraints updated along the learning process can produce faster convergence to adequate policies. The best results achieved present an important reduction over execution time (34,32%) and number of iterations (67.94%). This occurs due to the early state space reduction caused by shared information on state space constraints.
id ITA_8babd75d78d96c14f1166eebe3fd15a5
oai_identifier_str oai:agregador.ibict.br.BDTD_ITA:oai:ita.br:3096
network_acronym_str ITA
network_name_str Biblioteca Digital de Teses e Dissertações do ITA
spelling ERG-ARCH : a reinforcement learning architecture for propositionally constrained multi-agent state spacesAprendizagem (inteligência artificial)Sistemas tutores inteligentesProcessos de MarkovAvaliação de desempenho de softwareComputaçãoThe main goal of this work is to present an approach that ?nds an appropriate set of sequential actions for a group of cooperative agents interacting over a constrained environment. This search is considered a complex task for autonomous agents and is not possible to use default reinforcement learning algorithms to learn the adequate policy. In this thesis, a technique that deals with propositionally constrained state spaces and makes use of a Reinforcement Learning algorithm based on Markov Decision Process is proposed. A new model is also presented which formally de?nes this restricted search space. By so doing, this work aims at reducing the overall exploratory need, thus improving the performance of the learning algorithm. To constrain the state space the concept of extended reachability goals is employed. Through them it is possible to de?ne an objective to be preserved during the iteration with the environment and another that de?nes a goal state. In this cooperative environment, the information about the propositions is shared among the agents during its interaction. An architecture to solve problems in such environments is also presented. Experiments to validate the proposed algorithm were performed on different test cases and showed interesting results. A performance evaluation against standard Reinforcement Learning techniques showed that by extending autonomous learning with propositional constraints updated along the learning process can produce faster convergence to adequate policies. The best results achieved present an important reduction over execution time (34,32%) and number of iterations (67.94%). This occurs due to the early state space reduction caused by shared information on state space constraints.Instituto Tecnológico de AeronáuticaCarlos Henrique Costa RibeiroAnderson Viçoso de Araújo2014-10-06info:eu-repo/semantics/publishedVersioninfo:eu-repo/semantics/doctoralThesishttp://www.bd.bibl.ita.br/tde_busca/arquivo.php?codArquivo=3096reponame:Biblioteca Digital de Teses e Dissertações do ITAinstname:Instituto Tecnológico de Aeronáuticainstacron:ITAenginfo:eu-repo/semantics/openAccessapplication/pdf2019-02-02T14:05:05Zoai:agregador.ibict.br.BDTD_ITA:oai:ita.br:3096http://oai.bdtd.ibict.br/requestopendoar:null2020-05-28 19:41:06.68Biblioteca Digital de Teses e Dissertações do ITA - Instituto Tecnológico de Aeronáuticatrue
dc.title.none.fl_str_mv ERG-ARCH : a reinforcement learning architecture for propositionally constrained multi-agent state spaces
title ERG-ARCH : a reinforcement learning architecture for propositionally constrained multi-agent state spaces
spellingShingle ERG-ARCH : a reinforcement learning architecture for propositionally constrained multi-agent state spaces
Anderson Viçoso de Araújo
Aprendizagem (inteligência artificial)
Sistemas tutores inteligentes
Processos de Markov
Avaliação de desempenho de software
Computação
title_short ERG-ARCH : a reinforcement learning architecture for propositionally constrained multi-agent state spaces
title_full ERG-ARCH : a reinforcement learning architecture for propositionally constrained multi-agent state spaces
title_fullStr ERG-ARCH : a reinforcement learning architecture for propositionally constrained multi-agent state spaces
title_full_unstemmed ERG-ARCH : a reinforcement learning architecture for propositionally constrained multi-agent state spaces
title_sort ERG-ARCH : a reinforcement learning architecture for propositionally constrained multi-agent state spaces
author Anderson Viçoso de Araújo
author_facet Anderson Viçoso de Araújo
author_role author
dc.contributor.none.fl_str_mv Carlos Henrique Costa Ribeiro
dc.contributor.author.fl_str_mv Anderson Viçoso de Araújo
dc.subject.por.fl_str_mv Aprendizagem (inteligência artificial)
Sistemas tutores inteligentes
Processos de Markov
Avaliação de desempenho de software
Computação
topic Aprendizagem (inteligência artificial)
Sistemas tutores inteligentes
Processos de Markov
Avaliação de desempenho de software
Computação
dc.description.none.fl_txt_mv The main goal of this work is to present an approach that ?nds an appropriate set of sequential actions for a group of cooperative agents interacting over a constrained environment. This search is considered a complex task for autonomous agents and is not possible to use default reinforcement learning algorithms to learn the adequate policy. In this thesis, a technique that deals with propositionally constrained state spaces and makes use of a Reinforcement Learning algorithm based on Markov Decision Process is proposed. A new model is also presented which formally de?nes this restricted search space. By so doing, this work aims at reducing the overall exploratory need, thus improving the performance of the learning algorithm. To constrain the state space the concept of extended reachability goals is employed. Through them it is possible to de?ne an objective to be preserved during the iteration with the environment and another that de?nes a goal state. In this cooperative environment, the information about the propositions is shared among the agents during its interaction. An architecture to solve problems in such environments is also presented. Experiments to validate the proposed algorithm were performed on different test cases and showed interesting results. A performance evaluation against standard Reinforcement Learning techniques showed that by extending autonomous learning with propositional constraints updated along the learning process can produce faster convergence to adequate policies. The best results achieved present an important reduction over execution time (34,32%) and number of iterations (67.94%). This occurs due to the early state space reduction caused by shared information on state space constraints.
description The main goal of this work is to present an approach that ?nds an appropriate set of sequential actions for a group of cooperative agents interacting over a constrained environment. This search is considered a complex task for autonomous agents and is not possible to use default reinforcement learning algorithms to learn the adequate policy. In this thesis, a technique that deals with propositionally constrained state spaces and makes use of a Reinforcement Learning algorithm based on Markov Decision Process is proposed. A new model is also presented which formally de?nes this restricted search space. By so doing, this work aims at reducing the overall exploratory need, thus improving the performance of the learning algorithm. To constrain the state space the concept of extended reachability goals is employed. Through them it is possible to de?ne an objective to be preserved during the iteration with the environment and another that de?nes a goal state. In this cooperative environment, the information about the propositions is shared among the agents during its interaction. An architecture to solve problems in such environments is also presented. Experiments to validate the proposed algorithm were performed on different test cases and showed interesting results. A performance evaluation against standard Reinforcement Learning techniques showed that by extending autonomous learning with propositional constraints updated along the learning process can produce faster convergence to adequate policies. The best results achieved present an important reduction over execution time (34,32%) and number of iterations (67.94%). This occurs due to the early state space reduction caused by shared information on state space constraints.
publishDate 2014
dc.date.none.fl_str_mv 2014-10-06
dc.type.driver.fl_str_mv info:eu-repo/semantics/publishedVersion
info:eu-repo/semantics/doctoralThesis
status_str publishedVersion
format doctoralThesis
dc.identifier.uri.fl_str_mv http://www.bd.bibl.ita.br/tde_busca/arquivo.php?codArquivo=3096
url http://www.bd.bibl.ita.br/tde_busca/arquivo.php?codArquivo=3096
dc.language.iso.fl_str_mv eng
language eng
dc.rights.driver.fl_str_mv info:eu-repo/semantics/openAccess
eu_rights_str_mv openAccess
dc.format.none.fl_str_mv application/pdf
dc.publisher.none.fl_str_mv Instituto Tecnológico de Aeronáutica
publisher.none.fl_str_mv Instituto Tecnológico de Aeronáutica
dc.source.none.fl_str_mv reponame:Biblioteca Digital de Teses e Dissertações do ITA
instname:Instituto Tecnológico de Aeronáutica
instacron:ITA
reponame_str Biblioteca Digital de Teses e Dissertações do ITA
collection Biblioteca Digital de Teses e Dissertações do ITA
instname_str Instituto Tecnológico de Aeronáutica
instacron_str ITA
institution ITA
repository.name.fl_str_mv Biblioteca Digital de Teses e Dissertações do ITA - Instituto Tecnológico de Aeronáutica
repository.mail.fl_str_mv
subject_por_txtF_mv Aprendizagem (inteligência artificial)
Sistemas tutores inteligentes
Processos de Markov
Avaliação de desempenho de software
Computação
_version_ 1706809295615557632