Multiagent Particle Environments is a lightweight framework for simulating multi-agent reinforcement learning tasks in a continuous observation space with discrete action settings. It was originally developed by OpenAI and used in the influential paper Multi-Agent Actor-Critic for Mixed Cooperative-Competitive Environments. The environment provides simple particle-based worlds with simulated physics, where agents can move, communicate, and interact with each other. Scenarios are designed to model cooperative, competitive, and mixed interactions among agents, making it useful for testing algorithms in multi-agent settings. The project includes built-in scenarios such as navigation to landmarks, cooperative tasks, and adversarial setups. Although archived, its concepts and code structure remain foundational for more advanced libraries like PettingZoo, which extended and maintained this environment.
Features
- Simulated particle-based multi-agent environments
- Continuous observation and discrete action space
- Modular design for defining new scenarios
- Physics-based interaction engine for agents and landmarks
- Built-in rendering for visualizing agent behaviors
- Example scenarios supporting cooperative and competitive tasks