This repository provides a configurable warehouse simulation environment and reproducible testbed for benchmarking reinforcement learning (RL) algorithms.
The environment models:
- stochastic item arrivals
- dynamic order generation
- constrained agent movement
- action masking (invalid actions filtered)
Built on top of this environment, the testbed enables:
- reproducible scenario generation across heterogeneous warehouse layouts
- evaluation under diverse operating conditions (from low-load to highly congested settings)
- dataset export (CSV) for offline analysis
- training and benchmarking with PPO / MaskablePPO
git clone <your-repo-url>
cd warehouse-testbed
pip install -r requirements.txtRun a simple random agent:
python run_example.pyThis will:
- initialize the environment
- render the warehouse
- sample valid actions
- print rewards
Main file:
environment/env_storehouse.py
Gym-style usage:
state, _ = env.reset()
state, reward, terminated, truncated, info = env.step(action)mask = env.get_available_actions()Enable with:
record_scenario=TrueThis generates:
*_config.csv→ layout and parameters*_steps.csv→ initial state + events
Useful for:
- reproducibility
- dataset generation
- benchmarking
Run:
python benchmark_pipeline.pyThis will:
- generate scenarios
- train RL agents
- evaluate performance
- save results in
results_*
Defined in:
environment/conf.json
You can modify layouts, probabilities and warehouse settings.
warehouse-testbed/
├── README.md
├── requirements.txt
├── benchmark_pipeline.py
├── run_example.py
├── environment/
│ ├── env_storehouse.py
│ └── conf.json
- RL benchmarking
- warehouse optimization
- action masking research
- dataset generation
MIT License