Showing 8 open source projects for "cooperative"

View related business solutions
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 1
    LiteMultiAgent

    LiteMultiAgent

    The Library for LLM-based multi-agent applications

    LiteMultiAgent is a lightweight and extensible multi-agent reinforcement learning (MARL) platform designed for rapid experimentation. It allows researchers to design and test coordination, competition, and collaboration scenarios in simulated environments.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    talos

    talos

    Hyperparameter Optimization for TensorFlow, Keras and PyTorch

    Talos radically changes the ordinary Keras, TensorFlow (tf.keras), and PyTorch workflow by fully automating hyperparameter tuning and model evaluation. Talos exposes Keras and TensorFlow (tf.keras) and PyTorch functionality entirely and there is no new syntax or templates to learn. Talos is made for data scientists and data engineers that want to remain in complete control of their TensorFlow (tf.keras) and PyTorch models, but are tired of mindless parameter hopping and confusing...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Claude Code Bridge

    Claude Code Bridge

    Real-time multi-AI collaboration: Claude, Codex & Gemini

    Claude Code Bridge is an open-source command-line tool designed to enable real-time collaboration between multiple AI coding assistants within a unified development environment. The system allows developers to coordinate interactions between models such as Claude, Codex, and Gemini so that they can work together on programming tasks. By maintaining persistent shared context between these models, the tool reduces redundant prompts and minimizes token usage while allowing each AI system to...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Multi-Agent Particle Envs

    Multi-Agent Particle Envs

    Code for a multi-agent particle environment used in a paper

    Multiagent Particle Environments is a lightweight framework for simulating multi-agent reinforcement learning tasks in a continuous observation space with discrete action settings. It was originally developed by OpenAI and used in the influential paper Multi-Agent Actor-Critic for Mixed Cooperative-Competitive Environments. The environment provides simple particle-based worlds with simulated physics, where agents can move, communicate, and interact with each other. Scenarios are designed to model cooperative, competitive, and mixed interactions among agents, making it useful for testing algorithms in multi-agent settings. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 5
    Google Research Football

    Google Research Football

    Check out the new game server

    Google Research Football is a reinforcement learning environment simulating soccer matches. It focuses on learning complex behaviors such as team collaboration and strategy formation in competitive settings.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 6
    Multi-Agent Emergence Environments

    Multi-Agent Emergence Environments

    Environment generation code for the paper "Emergent Tool Use"

    multi-agent-emergence-environments is an open source research environment framework developed by OpenAI for the study of emergent behaviors in multi-agent systems. It was designed for the experiments described in the paper and blog post “Emergent Tool Use from Multi-Agent Autocurricula”, which investigated how complex cooperative and competitive behaviors can evolve through self-play. The repository provides environment generation code that builds on the mujoco-worldgen package, enabling dynamic creation of simulated physical environments. Developers can construct custom environments by combining modular components such as Boxes, Ramps, and RandomWalls using a flexible layering approach that reduces code duplication. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    SMAC

    SMAC

    SMAC: The StarCraft Multi-Agent Challenge

    SMAC (StarCraft II Multi-Agent Challenge) is a benchmark environment for cooperative multi-agent reinforcement learning (MARL), based on real-time strategy (RTS) game scenarios in StarCraft II. It allows researchers to test algorithms where multiple units (agents) must collaborate to win battles against built-in game AI opponents. SMAC provides a controlled testbed for studying decentralized execution and centralized training paradigms in MARL.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 8
    Spyse is a software framework for building multi-agent systems. It allows Python developers to build distributed intelligent systems of multiple cooperative agents based on FIPA, OWL, SOA and many others. Spyse is designed for ease-of-use and fun.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB