Showing 3 open source projects for "new protocol adhoc"

View related business solutions
  • Software Testing Platform | Testeum Icon
    Software Testing Platform | Testeum

    Testeum is a Software Testing & User Test platform

    Tired of bugs and poor UX going unnoticed despite thorough internal testing? Testeum is the SaaS crowdtesting platform that connects mobile and web app creators with carefully selected testers based on your criteria.
    Learn More
  • Top-Rated Free CRM Software Icon
    Top-Rated Free CRM Software

    216,000+ customers in over 135 countries grow their businesses with HubSpot

    HubSpot is an AI-powered customer platform with all the software, integrations, and resources you need to connect your marketing, sales, and customer service. HubSpot's connected platform enables you to grow your business faster by focusing on what matters most: your customers.
    Get started free
  • 1
    Elyra

    Elyra

    Elyra extends JupyterLab with an AI centric approach

    Elyra is a set of AI-centric extensions to JupyterLab Notebooks. The Elyra Getting Started Guide includes more details on these features. A version-specific summary of new features is located on the releases page.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    Rhasspy

    Rhasspy

    Offline private voice assistant for many human languages

    Rhasspy (ˈɹæspi) is an open-source, fully offline set of voice assistant services for many human languages that works well with Hermes protocol-compatible services (Snips.AI), Home Assistant and Hass.io, Node-RED, Jeedom, OpenHAB. Rhasspy will produce JSON events that can trigger action in home automation software, such as a Node-RED flow. Rhasspy comes with a snazzy web interface that lets you configure, program, and test your voice assistant remotely from your web browser. All of the web UI's...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 3
    EvalAI

    EvalAI

    Evaluating state of the art in AI

    ... add their own cluster of worker nodes to process participant submissions while we take care of hosting the challenge, handling user submissions, and maintaining the leaderboard. EvalAI lets participants submit code for their agent in the form of docker images which are evaluated against test environments on the evaluation server. During the evaluation, the worker fetches the image, test environment, and model snapshot and spins up a new container to perform the evaluation.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next