Classes | |
class | Acrobot |
Implementation of Acrobot game. More... | |
class | AggregatedPolicy |
class | AsyncLearning |
Wrapper of various asynchronous learning algorithms, e.g. More... | |
class | CartPole |
Implementation of Cart Pole task. More... | |
class | CategoricalDQN |
Implementation of the Categorical Deep Q-Learning network. More... | |
class | ContinuousActionEnv |
To use the dummy environment, one may start by specifying the state and action dimensions. More... | |
class | ContinuousDoublePoleCart |
Implementation of Continuous Double Pole Cart Balancing task. More... | |
class | ContinuousMountainCar |
Implementation of Continuous Mountain Car task. More... | |
class | DiscreteActionEnv |
To use the dummy environment, one may start by specifying the state and action dimensions. More... | |
class | DoublePoleCart |
Implementation of Double Pole Cart Balancing task. More... | |
class | DuelingDQN |
Implementation of the Dueling Deep Q-Learning network. More... | |
class | GreedyPolicy |
Implementation for epsilon greedy policy. More... | |
class | MountainCar |
Implementation of Mountain Car task. More... | |
class | NStepQLearningWorker |
Forward declaration of NStepQLearningWorker. More... | |
class | OneStepQLearningWorker |
Forward declaration of OneStepQLearningWorker. More... | |
class | OneStepSarsaWorker |
Forward declaration of OneStepSarsaWorker. More... | |
class | Pendulum |
Implementation of Pendulum task. More... | |
class | PrioritizedReplay |
Implementation of prioritized experience replay. More... | |
class | QLearning |
Implementation of various Q-Learning algorithms, such as DQN, double DQN. More... | |
class | RandomReplay |
Implementation of random experience replay. More... | |
class | RewardClipping |
Interface for clipping the reward to some value between the specified maximum and minimum value (Clipping here is implemented as .) More... | |
class | SAC |
Implementation of Soft Actor-Critic, a model-free off-policy actor-critic based deep reinforcement learning algorithm. More... | |
class | SimpleDQN |
class | SumTree |
Implementation of SumTree. More... | |
class | TrainingConfig |
Typedefs | |
template < typename EnvironmentType , typename NetworkType , typename UpdaterType , typename PolicyType > | |
using | NStepQLearning = AsyncLearning< NStepQLearningWorker< EnvironmentType, NetworkType, UpdaterType, PolicyType >, EnvironmentType, NetworkType, UpdaterType, PolicyType > |
Convenient typedef for async n step q-learning. More... | |
template < typename EnvironmentType , typename NetworkType , typename UpdaterType , typename PolicyType > | |
using | OneStepQLearning = AsyncLearning< OneStepQLearningWorker< EnvironmentType, NetworkType, UpdaterType, PolicyType >, EnvironmentType, NetworkType, UpdaterType, PolicyType > |
Convenient typedef for async one step q-learning. More... | |
template < typename EnvironmentType , typename NetworkType , typename UpdaterType , typename PolicyType > | |
using | OneStepSarsa = AsyncLearning< OneStepSarsaWorker< EnvironmentType, NetworkType, UpdaterType, PolicyType >, EnvironmentType, NetworkType, UpdaterType, PolicyType > |
Convenient typedef for async one step Sarsa. More... | |
using NStepQLearning = AsyncLearning<NStepQLearningWorker<EnvironmentType, NetworkType, UpdaterType, PolicyType>, EnvironmentType, NetworkType, UpdaterType, PolicyType> |
Convenient typedef for async n step q-learning.
EnvironmentType | The type of the reinforcement learning task. |
NetworkType | The type of the network model. |
UpdaterType | The type of the optimizer. |
PolicyType | The type of the behavior policy. |
Definition at line 233 of file async_learning.hpp.
using OneStepQLearning = AsyncLearning<OneStepQLearningWorker<EnvironmentType, NetworkType, UpdaterType, PolicyType>, EnvironmentType, NetworkType, UpdaterType, PolicyType> |
Convenient typedef for async one step q-learning.
EnvironmentType | The type of the reinforcement learning task. |
NetworkType | The type of the network model. |
UpdaterType | The type of the optimizer. |
PolicyType | The type of the behavior policy. |
Definition at line 197 of file async_learning.hpp.
using OneStepSarsa = AsyncLearning<OneStepSarsaWorker<EnvironmentType, NetworkType, UpdaterType, PolicyType>, EnvironmentType, NetworkType, UpdaterType, PolicyType> |
Convenient typedef for async one step Sarsa.
EnvironmentType | The type of the reinforcement learning task. |
NetworkType | The type of the network model. |
UpdaterType | The type of the optimizer. |
PolicyType | The type of the behavior policy. |
Definition at line 215 of file async_learning.hpp.