Showing 258 open source projects for "vb6 runtime plus"

View related business solutions
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build, govern, and optimize agents and models with Gemini Enterprise Agent Platform.
    Start Free
  • Compliant and Reliable File Transfers Backed by Top Security Certifications Icon
    Compliant and Reliable File Transfers Backed by Top Security Certifications

    Cerberus FTP Server delivers SOC 2 Type II certified security and FIPS 140-2 validated encryption.

    Stop relying on non-certified, legacy file transfer tools that creak under the weight of modern security demands. Get full audit trails, advanced access controls and more supported by an award-winning team of experts. Start your free 25-day trial today.
    Start Free Trial
  • 1
    ONNX Runtime

    ONNX Runtime

    ONNX Runtime: cross-platform, high performance ML inferencing

    ONNX Runtime is a cross-platform inference and training machine-learning accelerator. ONNX Runtime inference can enable faster customer experiences and lower costs, supporting models from deep learning frameworks such as PyTorch and TensorFlow/Keras as well as classical machine learning libraries such as scikit-learn, LightGBM, XGBoost, etc. ONNX Runtime is compatible with different hardware, drivers, and operating systems, and provides optimal performance by leveraging hardware accelerators...
    Downloads: 62 This Week
    Last Update:
    See Project
  • 2
    Android Runtime for NativeScript

    Android Runtime for NativeScript

    NativeScript for Android using v8

    Contains the source code for the NativeScript's Android Runtime. NativeScript is a framework that enables developers to write truly native mobile applications for Android and iOS using JavaScript and CSS. Each mobile platform has its own ecosystem and offers completely different development tools and language(s) - Java for Android and Objective C (Swift) for iOS. In order to translate JavaScript code to the corresponding native APIs some kind of proxy mechanism is needed. This is exactly...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 3
    WasmEdge

    WasmEdge

    Extensible WebAssembly runtime for cloud native applications

    WasmEdge is a lightweight, high-performance, and extensible WebAssembly runtime. It is the fastest Wasm VM today. WasmEdge is an official sandbox project hosted by the CNCF. Its use cases include modern web application architectures (Isomorphic & Jamstack applications), microservices on the edge cloud, serverless SaaS APIs, embedded functions, smart contracts, and smart devices. The WasmEdge Runtime provides a well-defined execution sandbox for its contained WebAssembly bytecode program. The...
    Downloads: 75 This Week
    Last Update:
    See Project
  • 4
    Bun

    Bun

    Incredibly fast JavaScript runtime, bundler, test runner

    Develop, test, run, and bundle JavaScript & TypeScript projects—all with Bun. Bun is an all-in-one JavaScript runtime & toolkit designed for speed, complete with a bundler, test runner, and Node. js-compatible package manager. Bun is a new JavaScript runtime built from scratch to serve the modern JavaScript ecosystem. Bun starts fast and runs fast. It extends JavaScriptCore, the performance-minded JS engine built for Safari. As computing moves to the edge, this is critical. Bun provides a...
    Downloads: 31 This Week
    Last Update:
    See Project
  • Build Securely on AWS with Proven Frameworks Icon
    Build Securely on AWS with Proven Frameworks

    Lay a foundation for success with Tested Reference Architectures developed by Fortinet’s experts. Learn more in this white paper.

    Moving to the cloud brings new challenges. How can you manage a larger attack surface while ensuring great network performance? Turn to Fortinet’s Tested Reference Architectures, blueprints for designing and securing cloud environments built by cybersecurity experts. Learn more and explore use cases in this white paper.
    Download Now
  • 5
    Luau

    Luau

    Fast, small, safe, gradually typed embeddable scripting language

    It is designed to be backward compatible with Lua 5.1, as well as incorporating some features from future Lua releases, but also expands the feature set (most notably with type annotations). Luau is largely implemented from scratch, with the language runtime being a very heavily modified version of Lua 5.1 runtime, with a completely rewritten interpreter and other performance innovations. The runtime mostly preserves Lua 5.1 API, so existing bindings should be more or less compatible with a...
    Downloads: 30 This Week
    Last Update:
    See Project
  • 6
    Electrobun

    Electrobun

    Build ultra fast, tiny, and cross-platform desktop apps

    Electrobun is an experimental JavaScript runtime and bundling environment designed to streamline local development and application execution by combining fast tooling with a simplified developer experience. The project focuses on providing a cohesive workflow where building, running, and managing JavaScript applications can happen with minimal configuration and overhead. It is built with performance in mind, aiming to reduce the friction typically associated with modern web tooling stacks by...
    Downloads: 9 This Week
    Last Update:
    See Project
  • 7
    LiteRT

    LiteRT

    LiteRT is the new name for TensorFlow Lite (TFLite)

    LiteRT is an experimental, real-time inference runtime built by Google AI Edge to run lightweight ML models on edge devices with ultra-low latency. It focuses on delivering predictable and consistent performance for models used in time-critical applications like robotics, AR/VR, and IoT. LiteRT is designed to be hardware-agnostic, with minimal dependencies and tight control over execution scheduling.
    Downloads: 15 This Week
    Last Update:
    See Project
  • 8
    Node.js

    Node.js

    Cross-platform JavaScript runtime environment

    Node.js is a free, open-source, cross-platform JavaScript runtime environment that allows developers to run JavaScript outside of the browser. It is built on Chrome’s V8 engine and is designed for building fast, scalable network applications. Node.js enables developers to create servers, APIs, command-line tools, and full-stack web applications using a single programming language. Its non-blocking, event-driven architecture makes it highly efficient for handling concurrent requests. The...
    Downloads: 19 This Week
    Last Update:
    See Project
  • 9
    IREE

    IREE

    A retargetable MLIR-based machine learning compiler runtime toolkit

    IREE (Intermediate Representation Execution Environment, pronounced as "eerie") is an MLIR-based end-to-end compiler and runtime that lowers Machine Learning (ML) models to a unified IR that scales up to meet the needs of the data center and down to satisfy the constraints and special considerations of mobile and edge deployments.
    Downloads: 8 This Week
    Last Update:
    See Project
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • 10
    OpenVR SDK

    OpenVR SDK

    API and runtime that allows access to VR hardware

    OpenVR is an API and runtime that allows access to VR hardware from multiple vendors without requiring that applications have specific knowledge of the hardware they are targeting. This repository is an SDK that contains the API and samples. The runtime is under SteamVR in Tools on Steam. The OpenVR API provides a game with a way to interact with Virtual Reality displays without relying on a specific hardware vendor's SDK. It can be updated independently of the game to add support for new...
    Downloads: 37 This Week
    Last Update:
    See Project
  • 11
    mlx

    mlx

    MLX: An array framework for Apple silicon

    MlX offers a local web interface to browse, download, and run ML models via Hugging Face or local sources. It supports searching by tags or tasks, visualization of model metadata, quick inference demos, automatic setup of runtime environments, and works with PyTorch, TensorFlow, and ONNX. Ideal for researchers exploring and testing models via browser.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 12
    Boxedwine

    Boxedwine

    Emulator that can run 32-bit Windows programs/games

    Boxedwine is a cross-platform emulator that enables users to run Windows applications by combining a Wine runtime with a custom emulated Linux kernel and CPU environment. Instead of relying on a native Linux system, it creates a fully self-contained execution environment that allows Windows binaries to run on platforms such as macOS, Linux, Windows, and even web browsers via WebAssembly. The project is written primarily in C++ and leverages SDL for graphics and input handling, providing a...
    Downloads: 9 This Week
    Last Update:
    See Project
  • 13
    DeepGEMM

    DeepGEMM

    Clean and efficient FP8 GEMM kernels with fine-grained scaling

    DeepGEMM is a specialized CUDA library for efficient, high-performance general matrix multiplication (GEMM) operations, with particular focus on low-precision formats such as FP8 (and experimental support for BF16). The library is designed to work cleanly and simply, avoiding overly templated or heavily abstracted code, while still delivering performance that rivals expert-tuned libraries. It supports both standard and “grouped” GEMMs, which is useful for architectures like Mixture of...
    Downloads: 107 This Week
    Last Update:
    See Project
  • 14
    DistroAV

    DistroAV

    DistroAV (formerly OBS-NDI): NDI integration for OBS Studio

    DistroAV is an open-source integration plugin for OBS Studio that provides Network Device Interface (NDI) support so users can send and receive live audio and video over IP networks directly within OBS, formerly known as the OBS-NDI project before being renamed. By implementing NDI input sources, dedicated output transports, and special filter modes, it allows creativity-oriented workflows such as capturing remote cameras, sharing scenes between machines, or distributing live feeds without...
    Downloads: 78 This Week
    Last Update:
    See Project
  • 15
    AsmJit

    AsmJit

    Low-latency machine code generation

    AsmJit is a low-level code generation library designed for dynamically creating machine code at runtime, enabling just-in-time (JIT) compilation for performance-critical applications. It provides a high-level API that abstracts away the complexity of writing raw assembly while still allowing fine-grained control over instruction generation. The library supports multiple architectures, including x86 and x64, making it versatile for cross-platform development. It is commonly used in...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    workerd

    workerd

    The JavaScript / Wasm runtime that powers Cloudflare Workers

    workerd is the server runtime that powers Cloudflare Workers — a modern, standards-aligned JavaScript and WebAssembly execution environment that can be run outside of Cloudflare’s global edge network as a standalone server or proxy. Designed to host serverless applications written to the Workers platform API (including fetch() and Web standard libraries), workerd enables developers to self-host edge-style workloads in local data centers or development environments, or to test Workers code...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 17
    Torch-TensorRT

    Torch-TensorRT

    PyTorch/TorchScript/FX compiler for NVIDIA GPUs using TensorRT

    Torch-TensorRT is a compiler for PyTorch/TorchScript, targeting NVIDIA GPUs via NVIDIA’s TensorRT Deep Learning Optimizer and Runtime. Unlike PyTorch’s Just-In-Time (JIT) compiler, Torch-TensorRT is an Ahead-of-Time (AOT) compiler, meaning that before you deploy your TorchScript code, you go through an explicit compile step to convert a standard TorchScript program into a module targeting a TensorRT engine. Torch-TensorRT operates as a PyTorch extension and compiles modules that integrate...
    Downloads: 15 This Week
    Last Update:
    See Project
  • 18
    AWS IoT Device SDK for C++ v2

    AWS IoT Device SDK for C++ v2

    Next generation AWS IoT Client SDK for C++ using AWS Common Runtime

    Next-generation AWS IoT Client SDK for C++ using the AWS common runtime. This document provides information about the AWS IoT device SDK for C++ V2. If you have any issues or feature requests, please file an issue or pull request. This SDK is built on the AWS Common Runtime, a collection of libraries (AWS-c-common, aws-c-io, aws-c-mqtt, aws-c-http, aws-c-cal, aws-c-auth, s2n...) written in C to be cross-platform, high-performance, secure, and reliable. The libraries are bound to C++ by the...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    CPU Features

    CPU Features

    A cross platform C99 library to get cpu features at runtime

    cpu_features is a cross-platform C library developed by Google that provides a simple and efficient way to detect available CPU features at runtime across a wide range of architectures and operating systems. It enables applications to determine which instruction sets (such as SSE, AVX, or NEON) are supported on the host machine, allowing developers to optimize performance dynamically. The library supports numerous architectures—including x86, ARM, AArch64, MIPS, POWER, RISCV, LoongArch, and...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 20
    CUDA API Wrappers

    CUDA API Wrappers

    Thin, unified, C++-flavored wrappers for the CUDA APIs

    CUDA API Wrappers is a C++ library providing high-level, modern wrappers for NVIDIA’s CUDA runtime and driver APIs, enhancing usability and efficiency. It is intended for those who would otherwise use these APIs directly, to make working with them more intuitive and consistent, making use of modern C++ language capabilities, programming idioms, and best practices. In a nutshell - making CUDA API work more fun.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 21
    whisper.cpp

    whisper.cpp

    Port of OpenAI's Whisper model in C/C++

    whisper.cpp is a lightweight, C/C++ reimplementation of OpenAI’s Whisper automatic speech recognition (ASR) model—designed for efficient, standalone transcription without external dependencies. The entire high-level implementation of the model is contained in whisper.h and whisper.cpp. The rest of the code is part of the ggml machine learning library. The command downloads the base.en model converted to custom ggml format and runs the inference on all .wav samples in the folder samples....
    Downloads: 378 This Week
    Last Update:
    See Project
  • 22
    CUDA-Q

    CUDA-Q

    C++ and Python support for the CUDA Quantum programming model

    CUDA-Q is an open-source platform for developing hybrid quantum-classical applications using a unified programming model across CPUs, GPUs, and quantum processing units. It provides a full toolchain that includes compilers, runtimes, and libraries for writing quantum programs in both C++ and Python. The platform is designed to be hardware-agnostic, allowing developers to run applications on different quantum backends or simulate them efficiently using GPU acceleration when physical quantum...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 23
    gVisor

    gVisor

    Application Kernel for Containers

    gVisor is an application kernel developed by Google that provides a strong layer of isolation between applications and the host operating system. Written in Go, it implements a Linux-compatible system call interface that runs entirely in user space, creating a secure sandboxed environment for containers. Unlike traditional virtual machines or lightweight syscall filters, gVisor follows a third approach that offers many of the security benefits of virtualization while maintaining the speed,...
    Downloads: 9 This Week
    Last Update:
    See Project
  • 24
    llama.cpp

    llama.cpp

    Port of Facebook's LLaMA model in C/C++

    The llama.cpp project enables the inference of Meta's LLaMA model (and other models) in pure C/C++ without requiring a Python runtime. It is designed for efficient and fast model execution, offering easy integration for applications needing LLM-based capabilities. The repository focuses on providing a highly optimized and portable implementation for running large language models directly within C/C++ environments.
    Downloads: 121 This Week
    Last Update:
    See Project
  • 25
    Google Highway

    Google Highway

    Performance-portable, length-agnostic SIMD with runtime dispatch

    Google Highway is a high-performance C++ library designed to provide portable SIMD (Single Instruction, Multiple Data) vectorization across multiple CPU architectures while maintaining predictable and efficient behavior. It abstracts low-level vector intrinsics into a consistent API that maps closely to hardware instructions, allowing developers to write high-performance code without relying heavily on compiler auto-vectorization. Highway enables the same source code to run across different...
    Downloads: 4 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • 5
  • Next
MongoDB Logo MongoDB