Skip to content
/ bomber Public

Reinforcement learning for bombing rocks in own environment

Notifications You must be signed in to change notification settings

eidelen/bomber

Repository files navigation

bomber test

bomber

Reinforcement learning project with own gym. This is the self chosen final assignment of an advanced machine learning course at IDSIA, SUPSI Lugano.

Bomber Environment

Given a simple 10 x 10 gridword, there is an agent surrounded by rocks. The main goal of the agent is to blast all the rocks. The agent's actions are moving up, down, left, right and detonate a bomb. The detonation destroys all rocks within a 3x3 square around the bomb, respectively the agent. The agent gets a reward for each blasted rock and a special reward when all rocks are destroyed. On the other side, the agent gets a penalty when colliding with the walls or rocks, when planting a bomb and for each move. All these environment properties can be customized.

The Simple Bomber

In the simple bomber case, the bomb blasts immediatly after the agent plants the bomb. Therefore the agent does not get a penalty when being close to the blasts. I tried many different varations of learning, like agent dies or gets penalty when colliding with rocks and wall.

The Reduced Simple Bomber

The reduced simple bomber has the same properties as the above simple bomber. But the reduced bomber agent only observes its surrounding fields - therefore the observation space is reduced to a patch of 3x3. This leads to a lack of knowledge for the agent, but also increases the learning speed and most importantly, one agent can be applied on different grid sizes.

Above agent was trained on a6x6 grid,but then also tested on an 8x8 and 10x10 play area. Unlike the simple bomber, the model here uses an additional LSTM layer. One can see that the agent isn't performing great on unknown grid sizes.

Next step was to train the agent with reduced observation space on different grid sizes, where the penalties and rewards have to be scaled related to the grid size. In particular, below agent was trained on randomly selected6x6 gridsand8x8 grids. This resulted in much better performance on unknown grid.

Even on ultra large grids the agent performs not bad. However, it suddenly makes a stupid mistake in an actual easy situation.

The Smart Bomber

In the smart bomber environment, a bomb is first placed and two steps later, the bomb explodes. The agent receives a penalty if it is within a certain range of the blast. Therefore the smart bomber agent escapes the blast after placing the bomb (see image below).

The learning performance was best when using a reduced observation space and a LSTM layer. Below examples is an agent which was trained on 6x6, 8x8 and 10x10 grid sizes.

Installation Notes

Windows (Python 3.9 and 3.10)

python -m venv venv
.\venv\Scripts\activate.ps1 # if problem: Open Powershell as Admin: > Set-ExecutionPolicy RemoteSigned
python -m pip install -U pip
python -m pip install -U setuptools
pip install torch torchvision
pip install -U "ray[rllib]"
pip install tqdm tensorboard matplotlib

OSX

python -m venv venv
source venv/bin/activate
python -m pip install -U pip
python -m pip install -U setuptools
pip install torch torchvision
pip install -U "ray[rllib]"
pip install tqdm tensorboard matplotlib

About

Reinforcement learning for bombing rocks in own environment

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages