Showing 6 open source projects for "calibrate"

View related business solutions
  • Build on Google Cloud with $300 in Free Credit Icon
    Build on Google Cloud with $300 in Free Credit

    New to Google Cloud? Get $300 in free credit to explore Compute Engine, BigQuery, Cloud Run, Vertex AI, and 150+ other products.

    Start your next project with $300 in free Google Cloud credit. Spin up VMs, run containers, query exabytes in BigQuery, or build AI apps with Vertex AI and Gemini. Once your credits are used, keep building with 20+ products with free monthly usage, including Compute Engine, Cloud Storage, GKE, and Cloud Run functions. Sign up to start building right away.
    Start Free Trial
  • Build AI Apps with Gemini 3 on Vertex AI Icon
    Build AI Apps with Gemini 3 on Vertex AI

    Access Google’s most capable multimodal models. Train, test, and deploy AI with 200+ foundation models on one platform.

    Vertex AI gives developers access to Gemini 3—Google’s most advanced reasoning and coding model—plus 200+ foundation models including Claude, Llama, and Gemma. Build generative AI apps with Vertex AI Studio, customize with fine-tuning, and deploy to production with enterprise-grade MLOps. New customers get $300 in free credits.
    Try Vertex AI Free
  • 1
    tslib

    tslib

    Touchscreen access library

    C library for filtering touchscreen events. tslib consists of the library libts and tools that help you calibrate and use it in your environment. tslib runs on various hardware architectures and operating systems, including GNU/Linux, FreeBSD, or Android/Linux. Apart from building the latest tarball release, running ./configure, make, and make install, tslib is available from distributors like Arch Linux / Arch Linux ARM, Buildroot, Debian / Ubuntu, Fedora, or OpenSUSE and their package management.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 2
    TensorRT

    TensorRT

    C++ library for high performance inference on NVIDIA GPUs

    ...TensorRT-based applications perform up to 40X faster than CPU-only platforms during inference. With TensorRT, you can optimize neural network models trained in all major frameworks, calibrate for lower precision with high accuracy, and deploy to hyperscale data centers, embedded, or automotive product platforms. TensorRT is built on CUDA®, NVIDIA’s parallel programming model, and enables you to optimize inference leveraging libraries, development tools, and technologies in CUDA-X™ for artificial intelligence, autonomous machines, high-performance computing, and graphics. ...
    Downloads: 25 This Week
    Last Update:
    See Project
  • 3
    TF Quant Finance

    TF Quant Finance

    High-performance TensorFlow library for quantitative finance

    ...It implements pricing engines, risk measures, stochastic models, optimizers, and random number generators that are differentiable and vectorized for accelerators. Users can value options and fixed-income instruments, simulate paths, fit curves, and calibrate models while leveraging TensorFlow’s jit compilation and automatic differentiation. The codebase is organized as modular math and finance primitives so you can combine building blocks or target end-to-end examples. It includes Bazel builds, tests, and example notebooks to accelerate learning and adoption in real workflows. With hardware acceleration and differentiable models, it enables modern techniques like gradient-based calibration and end-to-end learning of market dynamics.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Caméléon, source code

    Caméléon, source code

    a graphical language for functional programming

    Warning: This version is the source code of Caméléon, new web site will be released with tutorials & documentations. Caméléon Creator is a graphical algorithm editor designed for processing research, prototyping & calibration. With Caméléon, you construct & calibrate data processes with a simple drag & drop mechanism. Caméléon is a graphical language, based on the theory. Caméléon is a try to make data processing research & prototyping simple & addictive.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Easily Host LLMs and Web Apps on Cloud Run Icon
    Easily Host LLMs and Web Apps on Cloud Run

    Run everything from popular models with on-demand NVIDIA L4 GPUs to web apps without infrastructure management.

    Run frontend and backend services, batch jobs, host LLMs, and queue processing workloads without the need to manage infrastructure. Cloud Run gives you on-demand GPU access for hosting LLMs and running real-time AI—with 5-second cold starts and automatic scale-to-zero so you only pay for actual usage. New customers get $300 in free credit to start.
    Try Cloud Run Free
  • 5
    Music Master

    Music Master

    Tools recording, mixing, mastering and delivering music tracks

    Command line tool handling steps to clean, calibrate, process, mix, master and deliver music tracks from recordings. Easy-to-use configuration files drive the complete processes.
    Leader badge
    Downloads: 1 This Week
    Last Update:
    See Project
  • 6
    The project is intended to create a Beowulf cluster using Virtual machines. The application can be used to calibrate and estimate the hardware required to create a Beowulf Cluster.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB