Heuristically accelerated reinforcement learning modularization for multi-agent multi-objective problems
dc.contributor.author | Ferreira L.A. | |
dc.contributor.author | Costa Ribeiro C.H. | |
dc.contributor.author | Da Costa Bianchi R.A. | |
dc.date.accessioned | 2019-08-19T23:45:22Z | |
dc.date.available | 2019-08-19T23:45:22Z | |
dc.date.issued | 2014 | |
dc.identifier.citation | FERREIRA, LEONARDO ANJOLETTO; COSTA RIBEIRO, CARLOS HENRIQUE; DA COSTA BIANCHI, REINALDO AUGUSTO. Heuristically accelerated reinforcement learning modularization for multi-agent multi-objective problems. Applied Intelligence (Boston), v. 1, p. 1-12, 2014. | |
dc.identifier.issn | 0924-669X | |
dc.identifier.uri | https://repositorio.fei.edu.br/handle/FEI/1241 | |
dc.description.abstract | This article presents two new algorithms for finding the optimal solution of a Multi-agent Multi-objective Reinforcement Learning problem. Both algorithms make use of the concepts of modularization and acceleration by a heuristic function applied in standard Reinforcement Learning algorithms to simplify and speed up the learning process of an agent that learns in a multi-agent multi-objective environment. In order to verify performance of the proposed algorithms, we considered a predator-prey environment in which the learning agent plays the role of prey that must escape the pursuing predator while reaching for food in a fixed location. The results show that combining modularization and acceleration using a heuristics function indeed produced simplification and speeding up of the learning process in a complex problem when comparing with algorithms that do not make use of acceleration or modularization techniques, such as Q-Learning and Minimax-Q. © 2014 Springer Science+Business Media New York. | |
dc.relation.ispartof | Applied Intelligence | |
dc.rights | Acesso Restrito | |
dc.title | Heuristically accelerated reinforcement learning modularization for multi-agent multi-objective problems | |
dc.type | Artigo |
Files in this item
Files | Size | Format | View |
---|