• Stars
    star
    215
  • Rank 183,925 (Top 4 %)
  • Language
    Python
  • Created over 4 years ago
  • Updated 12 months ago

Reviews

There are no reviews yet. Be the first to send feedback to the community and the maintainers!

Repository Details

Reinforcement learning approach for job shop scheduling

A Reinforcement Learning Environment For Job-Shop Scheduling

This folder contains the implementation of the paper "A Reinforcement Learning Environment For Job-Shop Scheduling".

It contains the deep reinforcement learning approach we have developed to solve the Job-Shop Scheduling problem.

The optimized environment is available as a separate repository.

til

If you've found our work useful for your research, you can cite the paper as follows:

@misc{tassel2021reinforcement,
      title={A Reinforcement Learning Environment For Job-Shop Scheduling}, 
      author={Pierre Tassel and Martin Gebser and Konstantin Schekotihin},
      year={2021},
      eprint={2104.03760},
      archivePrefix={arXiv},
      primaryClass={cs.LG}
}

Getting Started

This code has been tested on Ubuntu 18.04 and MacOs 10.15. Some users have reported difficulties running this program on Windows.

This work uses Ray's RLLib, Tensorflow and Wandb.

Make sure you have git, cmake, zlib1g, and, on Linux, zlib1g-dev installed.

You also need to have a Weight and Bias account to log your metrics. Otherwise, just remove all occurrence of wandb and log the metrics in another way.

git clone https://github.com/prosysscience/JSS
cd JSS
pip install -r requirements.txt

Important: Your instance must follow Taillard's specification.

Project Organization

β”œβ”€β”€ README.md                 <- The top-level README for developers using this project.
└── JSS
    β”œβ”€β”€ dispatching_rules/      <- Contains the code to run the disptaching rule FIFO and MWTR.
    β”œβ”€β”€ instances/              <- All Taillard's instances + 5 Demirkol instances.
    β”œβ”€β”€ randomLoop/             <- A random loop with action mask, usefull to debug environment and
    |                             to check if our agent learn.
    β”œβ”€β”€ CP.py                   <- OR-Tool's cp model for the JSS problem.
    β”œβ”€β”€ CustomCallbacks.py      <- A special RLLib's callback used to save the best solution found.
    β”œβ”€β”€ default_config.py       <- default config used for the disptaching rules.
    β”œβ”€β”€ env_wrapper.py          <- Envrionment wrapper to save the action's of the best solution found
    β”œβ”€β”€ main.py                 <- PPO approach, the main file to call to reproduce our approach.
    └── models.py               <- Tensorflow model who mask logits of illegal actions.

License

MIT License