Abstract
We present an adaptation of the standard genetic
program (GP) to hierarchically decomposable,
multi-agent learning problems. To break down a problem
that requires cooperation of multiple agents, we use
the team objective function to derive a simpler,
intermediate objective function for pairs of
cooperating agents. We apply GP to optimize first for
the intermediate, then for the team objective function,
using the final population from the earlier GP as the
initial seed population for the next. This layered
learning approach facilitates the discovery of
primitive behaviors that can be reused and adapted
towards complex objectives based on a shared team goal.
We use this method to evolve agents to play a
subproblem of robotic soccer (keep-away soccer).
Finally, we show how layered learning GP evolves better
agents than standard GP, including GP with
automatically defined functions, and how the problem
decomposition results in a significant learning-speed
increase.
Users
Please
log in to take part in the discussion (add own reviews or comments).