26 projects for "kernel-ntfs" with 2 filters applied:

  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 1
    Kernel Memory

    Kernel Memory

    Research project. A Memory solution for users, teams, and applications

    Kernel Memory is an open-source reference architecture developed by Microsoft to help developers build memory systems for AI applications powered by large language models. The project focuses on enabling applications to store, index, and retrieve information so that AI systems can incorporate external knowledge when generating responses.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    tt-metal

    tt-metal

    TT-NN operator library, and TT-Metalium low level kernel programming

    tt-metal, also referred to in its documentation as TT-Metalium, is Tenstorrent’s low-level software development kit for programming applications on Tenstorrent AI accelerators. The project is designed for developers who need direct access to the company’s Tensix processor architecture, exposing a programming model that is closer to hardware control than high-level inference frameworks. Instead of following a traditional GPU model centered on massive thread parallelism, the platform is built...
    Downloads: 49 This Week
    Last Update:
    See Project
  • 3
    CUDA Agent

    CUDA Agent

    Large-Scale Agentic RL for High-Performance CUDA Kernel Generation

    CUDA Agent is a research-driven agentic reinforcement learning system designed to automatically generate and optimize high-performance CUDA kernels for GPU workloads. The project addresses the long-standing challenge that efficient CUDA programming typically requires deep hardware expertise by training an autonomous coding agent capable of iterative improvement through execution feedback. Its architecture combines large-scale data synthesis, a skill-augmented CUDA development environment,...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    ANE Training

    ANE Training

    Training neural networks on Apple Neural Engine via APIs

    ...It explores the internal software stack of the Apple Neural Engine by interfacing with private classes such as _ANEClient and compiling custom compute graphs in the MIL format. The project includes performance benchmarks and kernel breakdowns that show how different components of the training loop are distributed between the ANE and CPU. It is primarily intended as a research and educational proof of concept rather than a production library, highlighting what is technically possible with undocumented hardware access.
    Downloads: 0 This Week
    Last Update:
    See Project
  • AI-powered service management for IT and enterprise teams Icon
    AI-powered service management for IT and enterprise teams

    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity. Maximize operational efficiency with refreshingly simple, AI-powered Freshservice.
    Try it Free
  • 5
    DeepGEMM

    DeepGEMM

    Clean and efficient FP8 GEMM kernels with fine-grained scaling

    DeepGEMM is a specialized CUDA library for efficient, high-performance general matrix multiplication (GEMM) operations, with particular focus on low-precision formats such as FP8 (and experimental support for BF16). The library is designed to work cleanly and simply, avoiding overly templated or heavily abstracted code, while still delivering performance that rivals expert-tuned libraries. It supports both standard and “grouped” GEMMs, which is useful for architectures like Mixture of...
    Downloads: 33 This Week
    Last Update:
    See Project
  • 6
    DeepSeek-V3.2-Exp

    DeepSeek-V3.2-Exp

    An experimental version of DeepSeek model

    DeepSeek-V3.2-Exp is an experimental release of the DeepSeek model family, intended as a stepping stone toward the next generation architecture. The key innovation in this version is DeepSeek Sparse Attention (DSA), a sparse attention mechanism that aims to optimize training and inference efficiency in long-context settings without degrading output quality. According to the authors, they aligned the training setup of V3.2-Exp with V3.1-Terminus so that benchmark results remain largely...
    Downloads: 25 This Week
    Last Update:
    See Project
  • 7
    FlashMLA

    FlashMLA

    FlashMLA: Efficient Multi-head Latent Attention Kernels

    FlashMLA is a high-performance decoding kernel library designed especially for Multi-Head Latent Attention (MLA) workloads, targeting NVIDIA Hopper GPU architectures. It provides optimized kernels for MLA decoding, including support for variable-length sequences, helping reduce latency and increase throughput in model inference systems using that attention style. The library supports both BF16 and FP16 data types, and includes a paged KV cache implementation with a block size of 64 to efficiently manage memory during decoding. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    HunyuanImage-3.0

    HunyuanImage-3.0

    A Powerful Native Multimodal Model for Image Generation

    HunyuanImage-3.0 is a powerful, native multimodal text-to-image generation model released by Tencent’s Hunyuan team. It unifies multimodal understanding and generation in a single autoregressive framework, combining text and image modalities seamlessly rather than relying on separate image-only diffusion components. It uses a Mixture-of-Experts (MoE) architecture with many expert subnetworks to scale efficiently, deploying only a subset of experts per token, which allows large parameter...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 9
    OllamaSharp

    OllamaSharp

    The easiest way to use Ollama in .NET

    OllamaSharp is an open-source .NET library that provides strongly typed bindings for interacting with the Ollama API, making it easier for developers to integrate local large language models into C# and .NET applications. The project acts as a wrapper around the Ollama API, exposing all endpoints through asynchronous methods that allow developers to perform tasks such as generating text, creating embeddings, and managing models. It supports both local and remote Ollama instances, enabling...
    Downloads: 3 This Week
    Last Update:
    See Project
  • AI-generated apps that pass security review Icon
    AI-generated apps that pass security review

    Stop waiting on engineering. Build production-ready internal tools with AI—on your company data, in your cloud.

    Retool lets you generate dashboards, admin panels, and workflows directly on your data. Type something like “Build me a revenue dashboard on my Stripe data” and get a working app with security, permissions, and compliance built in from day one. Whether on our cloud or self-hosted, create the internal software your team needs without compromising enterprise standards or control.
    Try Retool free
  • 10
    how-to-optim-algorithm-in-cuda

    how-to-optim-algorithm-in-cuda

    How to optimize some algorithm in cuda

    how-to-optim-algorithm-in-cuda is an open educational repository focused on teaching developers how to optimize algorithms for high-performance execution on GPUs using CUDA. The project combines technical notes, code examples, and practical experiments that demonstrate how common computational kernels can be optimized to improve speed and memory efficiency. Instead of presenting only theoretical explanations, the repository includes hand-written CUDA implementations of fundamental operations...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    MemOS

    MemOS

    AI memory OS for LLM and Agent systems

    MemOS is an experimental operating system and runtime built around the concept of memory-centric computing, where memory objects are first-class citizens and program execution is organized around efficient, persistent memory access rather than traditional process and file system boundaries. The project explores rethinking system abstractions by tightly coupling computation with memory objects so that programs can operate on large datasets without expensive serialization or context switching....
    Downloads: 1 This Week
    Last Update:
    See Project
  • 12
    Deepnote

    Deepnote

    Deepnote is a drop-in replacement for Jupyter

    ...The project provides an AI-first computational environment where users can write, analyze, and share code, data, and visualizations in a single integrated workspace. Built on top of the Jupyter kernel ecosystem, it maintains compatibility with existing notebook workflows while introducing additional features focused on collaboration and automation. The system supports programming languages such as Python, R, and SQL and allows users to execute and analyze data directly within interactive notebooks. Deepnote emphasizes team-based data science by enabling real-time collaboration similar to shared document editors, allowing multiple users to work simultaneously on the same notebook environment.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    GPT-Code UI

    GPT-Code UI

    An open source implementation of OpenAI's ChatGPT Code interpreter

    An open source implementation of OpenAI's ChatGPT Code interpreter. Simply ask the OpenAI model to do something and it will generate & execute the code for you. You can put a .env in the working directory to load the OPENAI_API_KEY environment variable. For Azure OpenAI Services, there are also other configurable variables like deployment name. See .env.azure-example for more information. Note that model selection on the UI is currently not supported for Azure OpenAI Services.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    MLPACK is a C++ machine learning library with emphasis on scalability, speed, and ease-of-use. Its aim is to make machine learning possible for novice users by means of a simple, consistent API, while simultaneously exploiting C++ language features to provide maximum performance and flexibility for expert users. * More info + downloads: https://mlpack.org * Git repo: https://github.com/mlpack/mlpack
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    ConvNeXt

    ConvNeXt

    Code release for ConvNeXt model

    ConvNeXt is a modernized convolutional neural network (CNN) architecture designed to rival Vision Transformers (ViTs) in accuracy and scalability while retaining the simplicity and efficiency of CNNs. It revisits classic ResNet-style backbones through the lens of transformer design trends—large kernel sizes, inverted bottlenecks, layer normalization, and GELU activations—to bridge the performance gap between convolutions and attention-based models. ConvNeXt’s clean, hierarchical structure makes it efficient for both pretraining and fine-tuning across a wide range of visual recognition tasks. It achieves competitive or superior results on ImageNet and downstream datasets while being easier to deploy and train than transformers. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Pattern Recognition and Machine Learning

    Pattern Recognition and Machine Learning

    Repository of notes, code and notebooks in Python

    ...Each section of the repository corresponds to chapters in the book and includes code examples that demonstrate statistical modeling, machine learning methods, and Bayesian inference techniques. These notebooks provide visualizations and computational demonstrations that help clarify complex topics such as probabilistic models, neural networks, kernel methods, and graphical models. The repository also includes implementations of sampling methods, clustering algorithms, and dimensionality reduction techniques used throughout machine learning research.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Adaptive Gaussian Filtering

    Adaptive Gaussian Filtering

    Machine learning with Gaussian kernels.

    Libagf is a machine learning library that includes adaptive kernel density estimators using Gaussian kernels and k-nearest neighbours. Operations include statistical classification, interpolation/non-linear regression and pdf estimation. For statistical classification there is a borders training feature for creating fast and general pre-trained models that nonetheless return the conditional probabilities.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    An implementation of the Kernel-based Orthogonal Projections to Latent Structures (K-OPLS) method for MATLAB and R. The supplied functionality includes e.g. cross-validation, kernel parameter optimization, model diagnostics and plot tools.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    QVision: Computer Vision Library for Qt

    QVision: Computer Vision Library for Qt

    Computer vision and image processing library for Qt.

    ...Thanks to a large number of wrapper functions these objects can be used with highly efficient functionality from third party libraries such as OpenCV, GNU Scientific Library, Computational Geometry Algorithms Library, Intel's Math Kernel Library and Integrated Performance Primitives, the Octave library, etc...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 20
    The CommGen platform is intended to be a simple, scalable integration platform for small to medium problem sets. The architecture is simple, consisting of a distributed kernel, distributed O/S, agent and application layer.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    The Another Web Chat Bot (AWCB) is a kernel for easy web-based chat bot development. The current object-based version is designed to make the development of user's wcb modules simpler.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    Maiden-core is a Linux kernel robotics API module, extensible in Erc. The Maiden Project also documents and develops the Maiden architecture for mobile robots, a set of (optional) designs and standards for robots using the Maiden-core API.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Rescuecore is a java framework for developing Robocup Rescue agents. The framework provides functions for launching agents, connecting to the simulation kernel and sending/receiving commands from the kernel
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    A solution for Home automation, with HTTP access to manage Tini Board, X10 - I2C - LCD components. You can add your own manager and exchange his services with other. All manager are plug on a kernel (Servlet Java).
    Downloads: 2 This Week
    Last Update:
    See Project
  • 25
    This Meshula Kernel contains a reactive AI core, a high quality physics simulator, a math library, and assorted utility routines
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • Next
MongoDB Logo MongoDB