Showing 15 open source projects for "reduce"

View related business solutions
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 1
    DeepEP

    DeepEP

    DeepEP: an efficient expert-parallel communication library

    ...Because MoE architectures require routing inputs to different experts, communication overhead can become a bottleneck — DeepEP addresses that by providing optimized GPU kernels and efficient dispatch/combining logic. The library also supports low-precision operations (such as FP8) to reduce memory and bandwidth usage during communication. DeepEP is aimed at large-scale model inference or training systems where expert parallelism is used to scale model capacity without replicating entire networks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    Pysheeet

    Pysheeet

    Python Cheat Sheet

    Pysheeet is a community-driven collection of Python code snippets covering common patterns and tasks like sockets, file I/O, data structures, and more. Each snippet is concise and battle-tested, designed to save coding time and reduce boilerplate. With documentation hosted on Read the Docs and an active GitHub repo, it’s a go-to resource for Python developers.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 3
    Antigravity Awesome Skills

    Antigravity Awesome Skills

    The Ultimate Collection of 700+ Agentic Skills for Claude Code

    ...The project includes skill definitions, example prompts, and usage patterns that highlight how modular abilities can be assembled into functioning assistants. Because it aims to reduce cognitive overhead, many skills show how to structure intents, handle context, and orchestrate multi-step reasoning without deep technical complexity. It also serves as inspiration for users looking to prototype new use cases — from conversational helpers that answer questions to workflow automators that trigger actions.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 4
    SENAITE LIMS

    SENAITE LIMS

    SENAITE Meta Package

    ...SENAITE can be easily integrated with instruments by using off-the-shell interfaces for data import and export. Custom interfacing is supported too. Import instrument results and avoid human errors in the carrying-over process. Reduce the turnaround time on results report delivery. Assign priorities to samples and due dates for tests, plan and assign the daily work by using worksheets, and keep track of delayed tests immediately.
    Downloads: 0 This Week
    Last Update:
    See Project
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 5
    Avalanche

    Avalanche

    End-to-End Library for Continual Learning based on PyTorch

    Avalanche is an end-to-end Continual Learning library based on Pytorch, born within ContinualAI with the unique goal of providing a shared and collaborative open-source (MIT licensed) codebase for fast prototyping, training and reproducible evaluation of continual learning algorithms. Avalanche can help Continual Learning researchers in several ways. This module maintains a uniform API for data handling: mostly generating a stream of data from one or more datasets. It contains all the major...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    AIMET

    AIMET

    AIMET is a library that provides advanced quantization and compression

    Qualcomm Innovation Center (QuIC) is at the forefront of enabling low-power inference at the edge through its pioneering model-efficiency research. QuIC has a mission to help migrate the ecosystem toward fixed-point inference. With this goal, QuIC presents the AI Model Efficiency Toolkit (AIMET) - a library that provides advanced quantization and compression techniques for trained neural network models. AIMET enables neural networks to run more efficiently on fixed-point AI hardware...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 7
    fvcore

    fvcore

    Collection of common code shared among different research projects

    ...It provides numerics and loss layers (e.g., focal loss, smooth-L1, IoU/GIoU) implemented for speed and clarity, along with initialization helpers and normalization layers for building PyTorch models. Its common modules include timers, logging, checkpoints, registry patterns, and configuration helpers that reduce boilerplate in research code. A standout capability is FLOP and activation counting, which analyzes arbitrary PyTorch graphs to report cost by operator and by module for precise profiling. The file I/O layer (PathManager) abstracts local/remote storage so the same code can read from disks, cloud buckets, or HTTP endpoints. Because it is small, stable, and well-tested, fvcore is frequently imported by projects like Detectron2 and PyTorchVideo to avoid duplicating infrastructure and to keep research repos.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    DGL

    DGL

    Python package built to ease deep learning on graph

    Build your models with PyTorch, TensorFlow or Apache MXNet. Fast and memory-efficient message passing primitives for training Graph Neural Networks. Scale to giant graphs via multi-GPU acceleration and distributed training infrastructure. DGL empowers a variety of domain-specific projects including DGL-KE for learning large-scale knowledge graph embeddings, DGL-LifeSci for bioinformatics and cheminformatics, and many others. We are keen to bringing graphs closer to deep learning researchers....
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Groq Python

    Groq Python

    The official Python Library for the Groq API

    Groq Python is the official Python SDK for the Groq REST API, giving Python developers straightforward access to Groq’s LLM, chat, audio, and other AI services. Through this library, you can call Groq’s models from Python code — for example to request chat completions, code generation, transcription, or any supported endpoint — using idiomatic Python syntax. The SDK handles authentication (via environment variable or parameter), defines proper type-safe request/response data types, and...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 10
    databooks

    databooks

    A CLI tool to reduce the friction between data scientists

    databooks is a package to ease the collaboration between data scientists using Jupyter notebooks, by reducing the number of git conflicts between different notebooks and resolution of git conflicts when encountered. Simply specify the paths for notebook files to remove metadata. By doing so, we can already avoid many of the conflicts. Specify the paths for notebook files with conflicts to be fixed. Then, databooks finds the source notebooks that caused the conflicts and compares them (so no...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    FairScale

    FairScale

    PyTorch extensions for high performance and large scale training

    ...It introduced Fully Sharded Data Parallel (FSDP) style techniques that shard model parameters, gradients, and optimizer states across ranks to fit bigger models into the same memory budget. The library also provides pipeline parallelism, activation checkpointing, mixed precision, optimizer state sharding (OSS), and auto-wrapping policies that reduce boilerplate in complex distributed setups. Its components are modular, so teams can adopt just the sharding optimizer or the pipeline engine without rewriting their training loop. FairScale puts emphasis on correctness and debuggability, offering hook points, logging, and reference examples for common trainer patterns. Although many ideas have since landed in core PyTorch, FairScale remains a valuable reference and a practical toolbox for squeezing more performance out of multi-GPU and multi-node jobs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Name-That-Hash

    Name-That-Hash

    Identify MD5, SHA256 and 300+ other hashes

    Name-That-Hash is a modern hash identification system that tells you what type of hash you are looking at, supporting MD5, SHA-256, and more than 300 other hash types. It is designed as a successor and improvement to older tools like HashID and Hash-Identifier, focusing on up-to-date hash databases and better usability. One of its core ideas is popularity-aware ranking: when you feed in a hash, it prioritizes likely real-world types such as NTLM over obscure ones like Skype hashes, instead...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Sparse Attention

    Sparse Attention

    "Generating Long Sequences with Sparse Transformers" examples

    Sparse Attention is OpenAI’s code release for the Sparse Transformer model, introduced in the paper Generating Long Sequences with Sparse Transformers. It explores how modifying the self-attention mechanism with sparse patterns can reduce the quadratic scaling of standard transformers, making it possible to model much longer sequences efficiently. The repository provides implementations of sparse attention layers, training code, and evaluation scripts for benchmark datasets. It highlights both fixed and learnable sparsity patterns that trade off computational cost and model expressiveness. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 14
    Mixup-CIFAR10

    Mixup-CIFAR10

    mixup: Beyond Empirical Risk Minimization

    ...This repository implements mixup for the CIFAR-10 dataset, showcasing its effectiveness in improving generalization, stability, and calibration of neural networks. The approach acts as a regularizer, encouraging linear behavior in the feature space between samples, which helps reduce overfitting and enhance performance on unseen data.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    aeneas

    aeneas

    Automagically synchronize audio and text (aka forced alignment)

    aeneas is a Python/C library and a set of tools to automagically synchronize audio and text (aka forced alignment). aeneas automatically generates a synchronization map between a list of text fragments and an audio file containing the narration of the text. In computer science this task is known as (automatically computing a) forced alignment.
    Downloads: 8 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB