Skip to the content.

Home > Agents

rlai.core.Agent

Base class for all agents.

rlai.core.Human

An interactive, human-driven agent that prompts for actions at each time step.

rlai.core.MdpAgent

MDP agent. Adds the concepts of state, reward discounting, and policy-based action to the base agent.

rlai.core.StochasticMdpAgent

Stochastic MDP agent. Adds random selection of actions based on probabilities specified in the agent's policy.

rlai.core.environments.robocode.RobocodeAgent

Robocode agent.

rlai.core.environments.robocode_continuous_action.RobocodeAgent

Robocode agent.

rlai.gpi.state_action_value.ActionValueMdpAgent

A stochastic MDP agent whose policy is based on action-value estimation. This agent is generally appropriate for
    discrete and continuous state spaces in which we estimate the value of actions using tabular and
    function-approximation methods, respectively. The action space need to be discrete in all of these cases. If the
    action space is continuous, then consider the `ParameterizedMdpAgent`.

rlai.policy_gradient.ParameterizedMdpAgent

A stochastic MDP agent whose policy is directly parameterized. This agent is generally appropriate when both the
    state and action spaces are continuous. If the action space is discrete, then consider the `ActionValueMdpAgent`.