RL Shooter is a Python-based framework that integrates ViZDoom with OpenAI Gym APIs to create a flexible reinforcement learning environment for FPS games. Users can define custom scenarios, maps, and reward structures to train agents on navigation, target detection, and shooting tasks. With configurable observation frames, action spaces, and logging facilities, it supports popular deep RL libraries such as Stable Baselines and RLlib, enabling clear performance tracking and reproducibility across experiments.
Scalable MADDPG is a research-oriented framework for multi-agent reinforcement learning, offering a scalable implementation of the MADDPG algorithm. It features centralized critics during training and independent actors at runtime for stability and efficiency. The library includes Python scripts to define custom environments, configure network architectures, and adjust hyperparameters. Users can train multiple agents in parallel, monitor metrics, and visualize learning curves. It integrates with OpenAI Gym-like environments and supports GPU acceleration via TensorFlow. By providing modular components, Scalable MADDPG enables flexible experimentation on cooperative, competitive, or mixed multi-agent tasks, facilitating rapid prototyping and benchmarking.