Showing 3 open source projects for "robot framework test data"

View related business solutions
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build, govern, and optimize agents and models with Gemini Enterprise Agent Platform.
    Start Free
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    Let your crypto work for you

    Put idle assets to work with competitive interest rates, borrow without selling, and trade with precision. All in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 1
    TTRL

    TTRL

    Test-Time Reinforcement Learning

    TTRL is an open-source framework for test-time reinforcement learning in large language models, with a particular focus on reasoning tasks where ground-truth labels are not available during inference. The project addresses the problem of how to generate useful reward signals from unlabeled test-time data, and its central insight is that common test-time scaling practices such as majority voting can be repurposed into reward estimates for online reinforcement learning. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    ROSA

    ROSA

    I Agent designed to interact with ROS1- and ROS2-based robotics system

    ROSA, short for Robot Operating System Agent, is an AI-powered software assistant developed by NASA’s Jet Propulsion Laboratory to simplify interaction with robotic systems that use the Robot Operating System (ROS). The project provides a natural language interface that allows developers and operators to interact with robots by issuing commands or queries in conversational language. Built on top of frameworks such as LangChain and modern large language models, ROSA translates user...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    local-llm

    local-llm

    Run LLMs locally on Cloud Workstations

    ...The repository includes tools, Docker configurations, and command-line utilities that simplify the process of downloading, running, and interacting with language models directly on local or cloud-based workstations. This approach improves data privacy and control, as all inference can be performed locally without sending sensitive information to external APIs. It also integrates seamlessly with Google Cloud services, allowing developers to build and test AI-powered applications within the broader cloud ecosystem.
    Downloads: 1 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB