🚧👷🛑 Under Construction!!!
This repository contains an implementation of the Asynchronous Advantage Actor-Critic (A3C) algorithm using PyTorch. A3C is a reinforcement learning method that leverages parallel actor-learners to stabilize and speed up training, providing faster convergence and improved performance in complex environments. The algorithm is evaluated on various Atari environments using Gymnasium.
It's recommended to use a Conda environment to manage dependencies and avoid conflicts. You can create and activate a new Conda environment with the following commands:
conda create -n rl python=3.11
conda activate rlAfter activating the environment, install the required dependencies using:
pip install -r requirements.txtYou can run the A3C algorithm on any supported Gymnasium Atari environment with a discrete action space using the following command:
python main.py --env 'MsPacmanNoFrameskip-v4'-
Environment Selection: Use
-eor--envto specify the Gymnasium environment. The default isNone, so you must specify an environment.Example:
python main.py --env 'PongNoFrameskip-v4' -
Number of Training Episodes: Use
--n_gamesto specify the number of games the agent should play during training.Example:
python main.py --n_games 5000
-
Parallel Environments: Use
--n_envsto specify the number of parallel environments to run during training. The default is 4.Example:
python main.py --env 'AsterixNoFrameskip-v4' --n_envs 16
Using a Conda environment along with these flexible command-line options will help you efficiently manage your dependencies and customize the training process for your specific needs.
|
AirRaid
|
Alien
|
Amidar
|
|
|
|
|
Assault
|
Asterix
|
Asteroids
|
|
|
|
|
Atlantis
|
BankHeist
|
BattleZone
|
|
|
|
|
BeamRider
|
Berzerk
|
Bowling
|
|
|
|
|
Boxing
|
Breakout
|
Carnival
|
|
|
|
|
Centipede
|
ChopperCommand
|
CrazyClimber
|
|
|
|
Special thanks to Phil Tabor, an excellent teacher! I highly recommend his Youtube channel.



































