14 projects for "sensor" with 2 filters applied:

  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    Access competitive interest rates on your digital assets.

    Generate interest, borrow against your crypto, and trade a range of cryptocurrencies — all in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 1
    Seamless Communication

    Seamless Communication

    Foundational Models for State-of-the-Art Speech and Text Translation

    ...The motivation is to move beyond “text in, text out” and enable direct, live, multi-turn exchange involving language, gesture, gaze, vision, and modality switching without user friction. The system architecture includes a real-time multimodal signal pipeline for audio, video, and sensor data, a dialog manager that can decide when to act (speak, gesture, point) or query, and a cross-modal reasoning layer that fuses perception with semantic context. The research prototype includes components for visual grounding (understanding when a user references something in view), gesture recognition and synthesis, and turn-taking mechanisms that mirror human conversational timing. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    OSWorld

    OSWorld

    Benchmarking Multimodal Agents for Open-Ended Tasks

    OSWorld is an open-source synthetic world environment designed for embodied AI research and multi-agent learning. It provides a richly simulated 3D world where multiple agents can interact, perform tasks, and learn complex behaviors. OSWorld emphasizes multi-modal interaction, enabling agents to process visual, auditory, and symbolic data for grounded learning in a simulated world.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 3
    OpenPlanter

    OpenPlanter

    Language-model investigation agent with a terminal UI

    OpenPlanter is an open-source Python project focused on building an intelligent automated planting or gardening system powered by software control and data processing. The repository is designed to help developers and hobbyists create programmable plant management workflows that can monitor, schedule, and optimize growing conditions. It emphasizes automation and extensibility, allowing integration with sensors, environmental data, and control logic for smart cultivation setups. The system is...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    Autonomous Agents

    Autonomous Agents

    Autonomous Agents (LLMs) research papers. Updated Daily

    Autonomous-Agents is a research-focused repository that collects implementations, experiments, and academic resources related to autonomous multi-agent systems and intelligent robotics. The project explores how multiple agents can cooperate and interact with complex environments through machine learning, imitation learning, and multimodal sensing. It includes frameworks that integrate visual perception, tactile sensing, and spatial reasoning to guide the actions of robotic agents during...
    Downloads: 2 This Week
    Last Update:
    See Project
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 5
    PilottAI

    PilottAI

    Python framework for building scalable multi-agent systems

    pilottai is an AI-based autonomous drone navigation system utilizing reinforcement learning for real-time decision-making. It is designed for simulating and training drones to fly safely through dynamic environments using AI-based controllers.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    DriveLM

    DriveLM

    Driving with Graph Visual Question Answering

    ...The project introduces a new paradigm called graph visual question answering that structures reasoning about driving scenes through interconnected tasks such as perception, prediction, planning, and motion control. Instead of treating autonomous driving as a purely sensor-driven pipeline, DriveLM frames it as a reasoning problem where models answer structured questions about the environment to guide decision making. The system includes DriveLM-Data, a dataset built on driving environments such as nuScenes and CARLA, where human-written reasoning steps connect different layers of driving tasks. This design allows models to learn relationships between objects, behaviors, and navigation decisions through graph-structured logic.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    LSTMs for Human Activity Recognition

    LSTMs for Human Activity Recognition

    Human Activity Recognition example using TensorFlow on smartphone

    ...The repository includes data preprocessing scripts, neural network architecture definitions, and training pipelines that allow researchers to reproduce and modify the experiments. It serves as an educational example of how deep learning models can process temporal sensor signals for pattern recognition tasks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Activity Recognition

    Activity Recognition

    Resources about activity recognition

    ...The repository includes links to code in MATLAB, Python, summaries of algorithms, datasets, and relevant research papers. Feature extraction method summaries (e.g. motion, sensor, vision). Deep learning for activity recognition references.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9

    april

    Simulating worlds in a computer

    The April project defines a set of C++ objects (World, Actor, Sensor, Reflex, Brain, Actuator, Event) that help simulate various environments. The world has an amount of energy that the user may use to create Actors and components while Actors may also create other Actors from their own energy. The code is structured in several layers: - april-core simply defines the library and the objects - april-gui builds on the concepts in april-core, giving a visual interpretation to the concepts - AprilDream is a GUI program that uses april-gui objects to display a World in evolution. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Auth0 B2B Essentials: SSO, MFA, and RBAC Built In Icon
    Auth0 B2B Essentials: SSO, MFA, and RBAC Built In

    Unlimited organizations, 3 enterprise SSO connections, role-based access control, and pro MFA included. Dev and prod tenants out of the box.

    Auth0's B2B Essentials plan gives you everything you need to ship secure multi-tenant apps. Unlimited orgs, enterprise SSO, RBAC, audit log streaming, and higher auth and API limits included. Add on M2M tokens, enterprise MFA, or additional SSO connections as you scale.
    Sign Up Free
  • 10
    CommonReality aims to bring a standard to the cognitive architecture, robotics, and sensor development community. CommonReality specifies a common interface for parceling the world into cognitively processable components. See the forums and mailing list
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    TACS (Trust Ant Colony System) is a Trust model for P2P, Ad-hoc and Wireless Sensor networks (also valid for multi-agent systems) based on the bio-inspired algorithm ACS (Ant Colony System).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Container for robot/environment simulation. The container simulates an environment, and gives to the robots living in it access to sensor views. The goal of this project is to provide the open source community a tool for robot behavior test and challeng
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    A feedback-enhanced framework for routing in wireless sensor networks. Speeds up the process of implementing routing techniques. Enables piggybacking of routing information on data packets. Written for the Omnet++ simulator and the Mobility Framework.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    MEM Net - Mote EMulator Network. This project will focus on: 1) MEM - Wireless Sensor Node (mote) emulator 2) MEM Net - network of emulated motes So far, the only released package is “visual-sim-slides”. More comming next !
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB