ERG-ARCH : a reinforcement learning architecture for propositionally constrained multi-agent state spaces
| Main Author: | |
|---|---|
| Publication Date: | 2014 |
| Format: | Doctoral thesis |
| Language: | eng |
| Source: | Biblioteca Digital de Teses e Dissertações do ITA |
| Download full: | http://www.bd.bibl.ita.br/tde_busca/arquivo.php?codArquivo=3096 |
Summary: | The main goal of this work is to present an approach that ?nds an appropriate set of sequential actions for a group of cooperative agents interacting over a constrained environment. This search is considered a complex task for autonomous agents and is not possible to use default reinforcement learning algorithms to learn the adequate policy. In this thesis, a technique that deals with propositionally constrained state spaces and makes use of a Reinforcement Learning algorithm based on Markov Decision Process is proposed. A new model is also presented which formally de?nes this restricted search space. By so doing, this work aims at reducing the overall exploratory need, thus improving the performance of the learning algorithm. To constrain the state space the concept of extended reachability goals is employed. Through them it is possible to de?ne an objective to be preserved during the iteration with the environment and another that de?nes a goal state. In this cooperative environment, the information about the propositions is shared among the agents during its interaction. An architecture to solve problems in such environments is also presented. Experiments to validate the proposed algorithm were performed on different test cases and showed interesting results. A performance evaluation against standard Reinforcement Learning techniques showed that by extending autonomous learning with propositional constraints updated along the learning process can produce faster convergence to adequate policies. The best results achieved present an important reduction over execution time (34,32%) and number of iterations (67.94%). This occurs due to the early state space reduction caused by shared information on state space constraints. |
| id |
ITA_8babd75d78d96c14f1166eebe3fd15a5 |
|---|---|
| oai_identifier_str |
oai:agregador.ibict.br.BDTD_ITA:oai:ita.br:3096 |
| network_acronym_str |
ITA |
| network_name_str |
Biblioteca Digital de Teses e Dissertações do ITA |
| spelling |
ERG-ARCH : a reinforcement learning architecture for propositionally constrained multi-agent state spacesAprendizagem (inteligência artificial)Sistemas tutores inteligentesProcessos de MarkovAvaliação de desempenho de softwareComputaçãoThe main goal of this work is to present an approach that ?nds an appropriate set of sequential actions for a group of cooperative agents interacting over a constrained environment. This search is considered a complex task for autonomous agents and is not possible to use default reinforcement learning algorithms to learn the adequate policy. In this thesis, a technique that deals with propositionally constrained state spaces and makes use of a Reinforcement Learning algorithm based on Markov Decision Process is proposed. A new model is also presented which formally de?nes this restricted search space. By so doing, this work aims at reducing the overall exploratory need, thus improving the performance of the learning algorithm. To constrain the state space the concept of extended reachability goals is employed. Through them it is possible to de?ne an objective to be preserved during the iteration with the environment and another that de?nes a goal state. In this cooperative environment, the information about the propositions is shared among the agents during its interaction. An architecture to solve problems in such environments is also presented. Experiments to validate the proposed algorithm were performed on different test cases and showed interesting results. A performance evaluation against standard Reinforcement Learning techniques showed that by extending autonomous learning with propositional constraints updated along the learning process can produce faster convergence to adequate policies. The best results achieved present an important reduction over execution time (34,32%) and number of iterations (67.94%). This occurs due to the early state space reduction caused by shared information on state space constraints.Instituto Tecnológico de AeronáuticaCarlos Henrique Costa RibeiroAnderson Viçoso de Araújo2014-10-06info:eu-repo/semantics/publishedVersioninfo:eu-repo/semantics/doctoralThesishttp://www.bd.bibl.ita.br/tde_busca/arquivo.php?codArquivo=3096reponame:Biblioteca Digital de Teses e Dissertações do ITAinstname:Instituto Tecnológico de Aeronáuticainstacron:ITAenginfo:eu-repo/semantics/openAccessapplication/pdf2019-02-02T14:05:05Zoai:agregador.ibict.br.BDTD_ITA:oai:ita.br:3096http://oai.bdtd.ibict.br/requestopendoar:null2020-05-28 19:41:06.68Biblioteca Digital de Teses e Dissertações do ITA - Instituto Tecnológico de Aeronáuticatrue |
| dc.title.none.fl_str_mv |
ERG-ARCH : a reinforcement learning architecture for propositionally constrained multi-agent state spaces |
| title |
ERG-ARCH : a reinforcement learning architecture for propositionally constrained multi-agent state spaces |
| spellingShingle |
ERG-ARCH : a reinforcement learning architecture for propositionally constrained multi-agent state spaces Anderson Viçoso de Araújo Aprendizagem (inteligência artificial) Sistemas tutores inteligentes Processos de Markov Avaliação de desempenho de software Computação |
| title_short |
ERG-ARCH : a reinforcement learning architecture for propositionally constrained multi-agent state spaces |
| title_full |
ERG-ARCH : a reinforcement learning architecture for propositionally constrained multi-agent state spaces |
| title_fullStr |
ERG-ARCH : a reinforcement learning architecture for propositionally constrained multi-agent state spaces |
| title_full_unstemmed |
ERG-ARCH : a reinforcement learning architecture for propositionally constrained multi-agent state spaces |
| title_sort |
ERG-ARCH : a reinforcement learning architecture for propositionally constrained multi-agent state spaces |
| author |
Anderson Viçoso de Araújo |
| author_facet |
Anderson Viçoso de Araújo |
| author_role |
author |
| dc.contributor.none.fl_str_mv |
Carlos Henrique Costa Ribeiro |
| dc.contributor.author.fl_str_mv |
Anderson Viçoso de Araújo |
| dc.subject.por.fl_str_mv |
Aprendizagem (inteligência artificial) Sistemas tutores inteligentes Processos de Markov Avaliação de desempenho de software Computação |
| topic |
Aprendizagem (inteligência artificial) Sistemas tutores inteligentes Processos de Markov Avaliação de desempenho de software Computação |
| dc.description.none.fl_txt_mv |
The main goal of this work is to present an approach that ?nds an appropriate set of sequential actions for a group of cooperative agents interacting over a constrained environment. This search is considered a complex task for autonomous agents and is not possible to use default reinforcement learning algorithms to learn the adequate policy. In this thesis, a technique that deals with propositionally constrained state spaces and makes use of a Reinforcement Learning algorithm based on Markov Decision Process is proposed. A new model is also presented which formally de?nes this restricted search space. By so doing, this work aims at reducing the overall exploratory need, thus improving the performance of the learning algorithm. To constrain the state space the concept of extended reachability goals is employed. Through them it is possible to de?ne an objective to be preserved during the iteration with the environment and another that de?nes a goal state. In this cooperative environment, the information about the propositions is shared among the agents during its interaction. An architecture to solve problems in such environments is also presented. Experiments to validate the proposed algorithm were performed on different test cases and showed interesting results. A performance evaluation against standard Reinforcement Learning techniques showed that by extending autonomous learning with propositional constraints updated along the learning process can produce faster convergence to adequate policies. The best results achieved present an important reduction over execution time (34,32%) and number of iterations (67.94%). This occurs due to the early state space reduction caused by shared information on state space constraints. |
| description |
The main goal of this work is to present an approach that ?nds an appropriate set of sequential actions for a group of cooperative agents interacting over a constrained environment. This search is considered a complex task for autonomous agents and is not possible to use default reinforcement learning algorithms to learn the adequate policy. In this thesis, a technique that deals with propositionally constrained state spaces and makes use of a Reinforcement Learning algorithm based on Markov Decision Process is proposed. A new model is also presented which formally de?nes this restricted search space. By so doing, this work aims at reducing the overall exploratory need, thus improving the performance of the learning algorithm. To constrain the state space the concept of extended reachability goals is employed. Through them it is possible to de?ne an objective to be preserved during the iteration with the environment and another that de?nes a goal state. In this cooperative environment, the information about the propositions is shared among the agents during its interaction. An architecture to solve problems in such environments is also presented. Experiments to validate the proposed algorithm were performed on different test cases and showed interesting results. A performance evaluation against standard Reinforcement Learning techniques showed that by extending autonomous learning with propositional constraints updated along the learning process can produce faster convergence to adequate policies. The best results achieved present an important reduction over execution time (34,32%) and number of iterations (67.94%). This occurs due to the early state space reduction caused by shared information on state space constraints. |
| publishDate |
2014 |
| dc.date.none.fl_str_mv |
2014-10-06 |
| dc.type.driver.fl_str_mv |
info:eu-repo/semantics/publishedVersion info:eu-repo/semantics/doctoralThesis |
| status_str |
publishedVersion |
| format |
doctoralThesis |
| dc.identifier.uri.fl_str_mv |
http://www.bd.bibl.ita.br/tde_busca/arquivo.php?codArquivo=3096 |
| url |
http://www.bd.bibl.ita.br/tde_busca/arquivo.php?codArquivo=3096 |
| dc.language.iso.fl_str_mv |
eng |
| language |
eng |
| dc.rights.driver.fl_str_mv |
info:eu-repo/semantics/openAccess |
| eu_rights_str_mv |
openAccess |
| dc.format.none.fl_str_mv |
application/pdf |
| dc.publisher.none.fl_str_mv |
Instituto Tecnológico de Aeronáutica |
| publisher.none.fl_str_mv |
Instituto Tecnológico de Aeronáutica |
| dc.source.none.fl_str_mv |
reponame:Biblioteca Digital de Teses e Dissertações do ITA instname:Instituto Tecnológico de Aeronáutica instacron:ITA |
| reponame_str |
Biblioteca Digital de Teses e Dissertações do ITA |
| collection |
Biblioteca Digital de Teses e Dissertações do ITA |
| instname_str |
Instituto Tecnológico de Aeronáutica |
| instacron_str |
ITA |
| institution |
ITA |
| repository.name.fl_str_mv |
Biblioteca Digital de Teses e Dissertações do ITA - Instituto Tecnológico de Aeronáutica |
| repository.mail.fl_str_mv |
|
| subject_por_txtF_mv |
Aprendizagem (inteligência artificial) Sistemas tutores inteligentes Processos de Markov Avaliação de desempenho de software Computação |
| _version_ |
1706809295615557632 |