Showing 53 open source projects for "jit"

View related business solutions
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 1
    Criterium

    Criterium

    Benchmarking library for clojure

    Criterium is a robust benchmarking library for Clojure that addresses common statistical and JIT-related issues. It provides accurate timings through warm-up, garbage collection control, and statistical summaries—making microbenchmarking more reliable than using time. Statistical processing of multiple evaluations. Inclusion of a warm-up period, designed to allow the JIT compiler to optimise its code. Purging of gc before testing, to isolate timings from GC state prior to testing. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    Mercurius

    Mercurius

    Implement GraphQL servers and gateways with Fastify

    Mercurius is a GraphQL adapter for Fastify. Caching of query parsing and validation. Automatic loader integration to avoid 1 + N queries. Just-In-Time compiler via graphql-jit. Federation support via @mercuriusjs/federation, including Subscriptions. Gateway implementation via @mercuriusjs/gateway, including Subscriptions. Batched query support. Customisable persisted queries.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    XLS

    XLS

    XLS: Accelerated HW Synthesis

    ...At the front end, DSLX lets you describe algorithms with strong typing and familiar control flow while remaining synthesis-friendly. The compiler lowers DSLX into a rich intermediate representation, applies aggressive optimization and scheduling passes, and can either JIT the design for software simulation or emit Verilog for FPGA/ASIC flows. A key idea is “software-style” iteration: fast, deterministic simulation via the JIT encourages test-driven development and property checking before committing to RTL. XLS also provides tooling for pipelining, state insertion, and formal equivalence checks between different stages, giving developers confidence as designs evolve. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    Numba CUDA Target

    Numba CUDA Target

    The CUDA target for Numba

    Numba CUDA Target is NVIDIA’s maintained CUDA backend for the Numba JIT compiler, enabling developers to write GPU-accelerated code directly in Python. It allows users to define CUDA kernels using Python syntax, which are then compiled into efficient GPU code at runtime using LLVM-based toolchains. This approach significantly lowers the barrier to entry for GPU programming by eliminating the need to write CUDA C++ while still delivering high performance.
    Downloads: 3 This Week
    Last Update:
    See Project
  • Our Free Plans just got better! | Auth0 Icon
    Our Free Plans just got better! | Auth0

    With up to 25k MAUs and unlimited Okta connections, our Free Plan lets you focus on what you do best—building great apps.

    You asked, we delivered! Auth0 is excited to expand our Free and Paid plans to include more options so you can focus on building, deploying, and scaling applications without having to worry about your security. Auth0 now, thank yourself later.
    Try free now
  • 5
    Fury

    Fury

    A blazing fast multi-language serialization framework

    Fury is a blazing-fast multi-language serialization framework powered by jit(just-in-time compilation) and zero-copy, providing up to 170x performance and ultimate ease of use. Based on efficient JIT, struct deserialization can get 170x speed up compared to other serialization frameworks. In serialization scenarios, Fury can achieve a 100x speed up compared to traditional serialization frameworks. If you use Java clusters on a large scale, this will save a lot of computing resources.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Flax

    Flax

    Flax is a neural network library for JAX

    ...Its design separates pure computation from state by threading parameter collections and RNGs explicitly, enabling reproducibility, transformation, and easy experimentation with JAX transforms like jit, pmap, and vmap. Modules define parameterized computations, but initialization and application remain side-effect free, which pairs naturally with JAX’s staging and compilation model. Flax emphasizes composability: optimizers, training loops, and checkpointing are provided as examples or utilities rather than monolithic frameworks, encouraging research-friendly customization. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 7
    Tunix

    Tunix

    A JAX-native LLM Post-Training Library

    Tunix is a JAX-native library for post-training large language models, bringing supervised fine-tuning, reinforcement learning–based alignment, and knowledge distillation into one coherent toolkit. It embraces JAX’s strengths—functional programming, jit compilation, and effortless multi-device execution—so experiments scale from a single GPU to pods of TPUs with minimal code changes. The library is organized around modular pipelines for data loading, rollout, optimization, and evaluation, letting practitioners swap components without rewriting the whole stack. Examples and reference configs demonstrate end-to-end runs for common model families, helping teams reproduce baselines before customizing. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    MIR

    MIR

    A lightweight JIT compiler based on MIR

    MIR means Medium Internal Representation. MIR project goal is to provide a basis to implement fast and lightweight interpreters and JITs. Plans to try MIR light-weight JIT first for CRuby or/and MRuby implementation. Each module can contain functions and declarations and data. Each function has a signature (parameters and return types), local variables (including function arguments) and instructions. Each local variable has a type which can be only a 64-bit integer, float, double, or long double. Each instruction has an opcode and operands. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    CUDA Python

    CUDA Python

    Performance meets Productivity

    ...It acts as a metapackage composed of multiple submodules that provide both high-level and low-level access to CUDA functionality, including runtime APIs, driver APIs, and JIT compilation tools. The project is designed to simplify GPU programming by offering Pythonic abstractions while still exposing the full power of CUDA for advanced users. It integrates tightly with the broader Python GPU ecosystem, including Numba for kernel compilation and CCCL for parallel primitives, allowing developers to write performant code without leaving Python. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 10
    Proxy

    Proxy

    Proxy: Next Generation Polymorphism in C++

    ...There are many drawbacks in this mechanism, including life management (because each object may have a different size and ownership) and reflection (because it is hard to balance between usability and memory allocation). To workaround these drawbacks, some languages like Java or C# choose to sacrifice performance by introducing GC to facilitate lifetime management, and JIT-compile the source code at runtime to generate full metadata. We improved the theory and implemented it as a C++ library without sacrificing performance, proposing to merge it into the C++ standard.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    Numba

    Numba

    NumPy aware dynamic Python compiler using LLVM

    Numba is an open source JIT compiler that translates a subset of Python and NumPy code into fast machine code. Numba translates Python functions to optimized machine code at runtime using the industry-standard LLVM compiler library. Numba-compiled numerical algorithms in Python can approach the speeds of C or FORTRAN. You don't need to replace the Python interpreter, run a separate compilation step, or even have a C/C++ compiler installed.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 12
    RustPython

    RustPython

    A Python Interpreter written in Rust

    RustPython is an implementation of the Python programming language written in Rust. Its goal is to provide a fast, embeddable, and secure interpreter that can be integrated into Rust applications or used standalone. Because it’s in Rust, it benefits from memory safety and modern tooling, allowing developers to compile Python into efficient binaries or embed it as a scripting engine in Rust projects. The interpreter aims to support the Python standard library, dynamic typing, garbage...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 13
    LÖVR

    LÖVR

    Lua Virtual Reality engine

    ...Out of the box you get fast stereo rendering, VR controllers, hand tracking, haptics, spatial audio, 3D physics, PBR materials, SDF fonts, multiplayer, and more. The entire engine is 1MB and runs on LuaJIT, the fastest JIT compiler round these parts. Projects are just folders with scripts and assets in them, organized however you want.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 14
    RLax

    RLax

    Library of JAX-based building blocks for reinforcement learning agents

    ...Rather than implementing full algorithms, RLax focuses on the core functional operations that underpin RL methods—such as computing value functions, returns, policy gradients, and loss terms—allowing researchers to flexibly assemble their own agents. It supports both on-policy and off-policy learning, as well as value-based, policy-based, and model-based approaches. RLax is fully JIT-compilable with JAX, enabling high-performance execution across CPU, GPU, and TPU backends. The library implements tools for Bellman equations, return distributions, general value functions, and policy optimization in both continuous and discrete action spaces. It integrates seamlessly with DeepMind’s Haiku (for neural network definition) and Optax (for optimization), making it a key component in modular RL pipelines.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Shumai

    Shumai

    Fast Differentiable Tensor Library in JavaScript & TypeScript with Bun

    Shumai is an experimental differentiable tensor library for TypeScript and JavaScript, developed by Facebook Research. It provides a high-performance framework for numerical computing and machine learning within modern JavaScript runtimes. Built on Bun and Flashlight, with ArrayFire as its numerical backend, Shumai brings GPU-accelerated tensor operations, automatic differentiation, and scientific computing tools directly to JavaScript developers. It allows seamless integration of machine...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16

    Halide

    A language for fast, portable data-parallel computation

    ...It isn't a standalone programming language however; rather it is embedded in C++ which means that you write C++ code, building an in-memory representation of a Halide pipeline using Halide's C++ API. This representation can then be compiled to an object file, or a JIT-compile and run in the same process. Halide also comes with a Python binding, allowing the writing of Halide embedded in Python without C++.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Jittor

    Jittor

    Jittor is a high-performance deep learning framework

    Jittor is a high-performance deep learning framework based on JIT compiling and meta-operators. The whole framework and meta-operators are compiled just in time. A powerful op compiler and tuner are integrated into Jittor. It allowed us to generate high-performance code specialized for your model. Jittor also contains a wealth of high-performance model libraries, including image recognition, detection, segmentation, generation, differentiable rendering, geometric learning, reinforcement learning, etc. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    Python Outlier Detection

    Python Outlier Detection

    A Python toolbox for scalable outlier detection

    PyOD is a comprehensive and scalable Python toolkit for detecting outlying objects in multivariate data. This exciting yet challenging field is commonly referred as outlier detection or anomaly detection. PyOD includes more than 30 detection algorithms, from classical LOF (SIGMOD 2000) to the latest COPOD (ICDM 2020) and SUOD (MLSys 2021). Since 2017, PyOD [AZNL19] has been successfully used in numerous academic researches and commercial products [AZHC+21, AZNHL19]. PyOD has multiple neural...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    Objeck

    Objeck

    Simple ⇒ Object-Oriented ⇒ Functional

    Objeck is an object-oriented programming language with functional features. Objeck emphasizes, expression, simplicity, portability, and scalability. The programming environment consists of a compiler, virtual machine, and command line debugger with IDE plugins.
    Leader badge
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    FrenchKiss.js

    FrenchKiss.js

    The blazing fast lightweight internationalization (i18n) module

    ...It provides a simple and really fast solution for handling internationalization. FrenchKiss is by now, the fastest i18n JS package out there, working 5 to 1000 times faster than any others by JIT compiling the translations, try it by running the benchmarks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    PI-Based Image Encoder / Converter

    PI-Based Image Encoder / Converter

    Python code able to convert / compress image to PI (3.14, π) Indexes

    ...ZIP also include 16 MB file with 16,7 mil numbers of PI Benchmark(Single-Thread): Hardware & Environment Apple Silicon: Apple M2 (Mac mini/MacBook) x86_64 Platform: Intel Core Ultra 5 225F (Arrow Lake, 10 Cores) OS 1: Fedora 43 (GNOME) OS 2: Windows 11 Pro (23H2/24H2) Software: Python 3.14.3 + Numba JIT (latest) Results (Lower is better) Platform / OS CPU Time (Seconds) macOS (Native) Apple M2 52.151311 s (in default setup) Fedora Linux Intel Core Ultra 5 225F 58.536457 s (in default Power Management: Balanced) Windows 11 Intel Core Ultra 5 225F 59.681427 s (important! Power Management set as High Performance .. in Balanced is slow)
    Downloads: 6 This Week
    Last Update:
    See Project
  • 22
    ts-node

    ts-node

    TypeScript execution and REPL for node.js

    It JIT transforms TypeScript into JavaScript, enabling you to directly execute TypeScript on Node.js without precompiling. This is accomplished by hooking node's module loading APIs, enabling it to be used seamlessly alongside other Node.js tools and libraries. Automatic defaults to match your node version, integrate with test runners, debuggers, and CLI tools, compatible with pre-compilation for production.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 23
    MyJIT is a small library allowing to (JIT) compile intermediate language into the machine code and subsequently execute it.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    Ravi Programming Language

    Ravi Programming Language

    Ravi is a dialect of Lua, featuring limited optional static typing

    Ravi Programming Language is a derivative of Lua 5.3 with limited optional static typing and MIR-based JIT compiler. Ravi is a derivative/dialect of Lua with limited optional static typing and JIT/AOT compilers. The name Ravi comes from the Sanskrit word for the Sun. Lua is perfect as a small embeddable dynamic language so why a derivative? Ravi extends Lua with static typing for greater performance under JIT compilation. However, static typing is optional and therefore Lua 5.3 programs are also valid Ravi programs. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 25
    CoreCLR

    CoreCLR

    Runtime for .NET Core

    CoreCLR is the runtime for .NET Core, an open source, general purpose development platform maintained by Microsoft and the .NET community. It is a cross-platform runtime for cloud, IoT, and desktop apps and includes a just-in-time compiler called RyuJIT.
    Downloads: 6 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • Next
MongoDB Logo MongoDB