Showing 24 open source projects for "private-gpt"

View related business solutions
  • Fully Managed MySQL, PostgreSQL, and SQL Server Icon
    Fully Managed MySQL, PostgreSQL, and SQL Server

    Automatic backups, patching, replication, and failover. Focus on your app, not your database.

    Cloud SQL handles your database ops end to end, so you can focus on your app.
    Try Free
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 1
    OpenAI-API-dotnet

    OpenAI-API-dotnet

    An unofficial C#/.NET SDK for accessing the OpenAI GPT-3 API

    A simple C# .NET wrapper library to use with OpenAI's API. More context on my blog. This is my original unofficial wrapper library around the OpenAI API.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    PostgresML

    PostgresML

    The GPU-powered AI application database

    PostgresML is a complete platform in a PostgreSQL extension. Build simpler, faster, and more scalable models right inside your database. Explore the SDK and test open source models in our hosted database. Combine and automate the entire workflow from embedding generation to indexing and querying for the simplest (and fastest) knowledge-based chatbot implementation. Leverage multiple types of natural language processing and machine learning models such as vector search and personalization...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    rust-bert

    rust-bert

    Rust native ready-to-use NLP pipelines and transformer-based models

    ...The project ports many capabilities of the Hugging Face Transformers ecosystem into the Rust programming language. It allows developers to run state-of-the-art NLP models like BERT, GPT-2, and DistilBERT directly within Rust applications while maintaining high performance and memory efficiency. The library integrates with Rust machine learning infrastructure using crates such as tch-rs and ONNX Runtime for model execution. It also includes tokenization utilities, model architectures, and task-specific pipelines that simplify the development of NLP applications. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    OpenBB

    OpenBB

    Investment Research for Everyone, Everywhere

    ...Customize your dashboards to build your dream terminal, integrate with your private datasets and bring your own fine-tuned AI copilots.
    Downloads: 9 This Week
    Last Update:
    See Project
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 5
    ANE Training

    ANE Training

    Training neural networks on Apple Neural Engine via APIs

    ANE Training is an experimental research project that demonstrates how to train neural networks directly on Apple’s Neural Engine by leveraging reverse-engineered private APIs that are normally inaccessible to developers. The repository implements a from-scratch transformer training pipeline capable of running both forward and backward passes on ANE hardware without relying on CoreML, Metal, or GPU acceleration. It explores the internal software stack of the Apple Neural Engine by interfacing with private classes such as _ANEClient and compiling custom compute graphs in the MIL format. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    TorchCode

    TorchCode

    Practice implementing softmax, attention, GPT-2 and more

    TorchCode is an interactive learning and practice platform designed to help developers master PyTorch by implementing core machine learning operations and architectures from scratch. It is structured similarly to competitive programming platforms like LeetCode but focuses specifically on tensor operations and deep learning concepts. The platform provides a collection of curated problems that cover fundamental topics such as activation functions, normalization layers, attention mechanisms,...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 7
    Porcupine

    Porcupine

    On-device wake word detection powered by deep learning

    Build always-listening yet private voice applications. Porcupine is a highly-accurate and lightweight wake word engine. It enables building always-listening voice-enabled applications. It is using deep neural networks trained in real-world environments. Compact and computationally-efficient. It is perfect for IoT. Cross-platform. Arm Cortex-M, STM32, PSoC, Arduino, and i.MX RT.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 8
    Transformer Engine

    Transformer Engine

    A library for accelerating Transformer models on NVIDIA GPUs

    ...TE also includes a framework-agnostic C++ API that can be integrated with other deep-learning libraries to enable FP8 support for Transformers. As the number of parameters in Transformer models continues to grow, training and inference for architectures such as BERT, GPT, and T5 become very memory and compute-intensive. Most deep learning frameworks train with FP32 by default. This is not essential, however, to achieve full accuracy for many deep learning models.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 9
    x-transformers

    x-transformers

    A simple but complete full-attention transformer

    A simple but complete full-attention transformer with a set of promising experimental features from various papers. Proposes adding learned memory key/values prior to attending. They were able to remove feedforwards altogether and attain a similar performance to the original transformers. I have found that keeping the feedforwards and adding the memory key/values leads to even better performance. Proposes adding learned tokens, akin to CLS tokens, named memory tokens, that is passed through...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 10
    Colossal-AI

    Colossal-AI

    Making large AI models cheaper, faster and more accessible

    ...This imposes challenges to the memory wall of the current accelerator hardware such as GPU. It is never ideal to train large models such as Vision Transformer, BERT, and GPT on a single GPU or a single machine. There is an urgent demand to train models in a distributed environment. However, distributed training, especially model parallelism, often requires domain expertise in computer systems and architecture. It remains a challenge for AI researchers to implement complex distributed training solutions for their models. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    Core ML Tools

    Core ML Tools

    Core ML tools contain supporting tools for Core ML model conversion

    ...Core ML optimizes on-device performance by leveraging the CPU, GPU, and Neural Engine while minimizing its memory footprint and power consumption. Running a model strictly on the user’s device removes any need for a network connection, which helps keep the user’s data private and your app responsive.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 12
    The Machine & Deep Learning Compendium

    The Machine & Deep Learning Compendium

    List of references in my private & single document

    The Machine & Deep Learning Compendium is an open-source knowledge repository that compiles summaries, references, and learning materials related to machine learning and deep learning. The project functions as a comprehensive compendium that organizes hundreds of topics covering algorithms, frameworks, research areas, and practical machine learning workflows. Originally created as a personal knowledge base, the repository evolved into a public educational resource designed to help learners...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    OneFlow

    OneFlow

    OneFlow is a deep learning framework designed to be user-friendly

    ...Provides a variety of services from primary AI talent training to enterprise-level machine learning lifecycle integrated management (MLOps), including AI training and AI development, and supports three deployment modes of public cloud, private cloud and hybrid cloud.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 14
    Promptify

    Promptify

    se GPT or other prompt based models to get structured output

    Promptify is an open-source Python library designed to simplify prompt engineering and the development of natural language processing pipelines using large language models. The project provides tools that help developers generate structured prompts for different NLP tasks and apply them across multiple generative AI systems. Instead of manually crafting prompts for each task, Promptify introduces a unified architecture that combines prompt templates, language model interfaces, and processing...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Transformer Reinforcement Learning X

    Transformer Reinforcement Learning X

    A repo for distributed training of language models with Reinforcement

    ...Training support for Hugging Face models is provided by Accelerate-backed trainers, allowing users to fine-tune causal and T5-based language models of up to 20B parameters, such as facebook/opt-6.7b, EleutherAI/gpt-neox-20b, and google/flan-t5-xxl. For models beyond 20B parameters, trlX provides NVIDIA NeMo-backed trainers that leverage efficient parallelism techniques to scale effectively.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    picoGPT

    picoGPT

    An unnecessarily tiny implementation of GPT-2 in NumPy

    picoGPT is a minimal implementation of the GPT-2 language model designed to demonstrate how transformer-based language models work at a conceptual level. The repository focuses on educational clarity rather than production performance, implementing the core components of the GPT architecture in a concise and readable way. It allows users to understand how tokenization, transformer layers, attention mechanisms, and autoregressive text generation operate in modern large language models. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    DialoGPT

    DialoGPT

    Large-scale pretraining for dialogue

    DialoGPT is an open-source conversational language model developed by Microsoft Research for generating natural dialogue responses using large-scale transformer architectures. The system is built on the GPT-2 architecture and is designed specifically for multi-turn conversation tasks, enabling machines to produce coherent responses during interactive dialogue. The model was trained on a massive dataset of approximately 147 million conversational exchanges extracted from Reddit discussion threads, allowing it to learn patterns of natural human conversation. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    KoboldAI

    KoboldAI

    Your gateway to GPT writing

    This is a browser-based front-end for AI-assisted writing with multiple local & remote AI models. It offers the standard array of tools, including Memory, Author's Note, World Info, Save & Load, adjustable AI settings, formatting options, and the ability to import existing AI Dungeon adventures. You can also turn on Adventure mode and play the game like AI Dungeon Unleashed. Stories can be played like a Novel, a text adventure game or used as a chatbot with an easy toggles to change...
    Leader badge
    Downloads: 172 This Week
    Last Update:
    See Project
  • 19
    PromptSource

    PromptSource

    Toolkit for creating, sharing and using natural language prompts

    PromptSource is a toolkit for creating, sharing and using natural language prompts. Recent work has shown that large language models exhibit the ability to perform reasonable zero-shot generalization to new tasks. For instance, GPT-3 demonstrated that large language models have strong zero- and few-shot abilities. FLAN and T0 then demonstrated that pre-trained language models fine-tuned in a massively multitask fashion yield even stronger zero-shot performance. A common denominator in these works is the use of prompts which has gained interest among NLP researchers and engineers. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    The fastai book

    The fastai book

    The fastai book, published as Jupyter Notebooks

    ...The remainder (including all markdown cells in the notebooks and other prose) is not licensed for any redistribution or change of format or medium, other than making copies of the notebooks or forking this repo for your own private use. No commercial or broadcast use is allowed. We are making these materials freely available to help you learn deep learning, so please respect our copyright and these restrictions.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    TensorFlow 2.0 Tutorials

    TensorFlow 2.0 Tutorials

    TensorFlow 2.x version's Tutorials and Examples

    ...These examples cover a wide range of topics including convolutional neural networks, recurrent neural networks, generative adversarial networks, autoencoders, and transformer-based models such as GPT and BERT. Each section of the repository includes runnable code and structured experiments designed to illustrate how different architectures and algorithms function in real applications. The tutorials use well-known benchmark datasets such as MNIST, CIFAR, and Fashion-MNIST to demonstrate practical model training and evaluation workflows.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    CrypTen

    CrypTen

    A framework for Privacy Preserving Machine Learning

    ...The framework supports both encryption and decryption of tensors and operations such as addition and multiplication over encrypted values. Although not yet production-ready, CrypTen focuses on advancing real-world secure ML applications, such as training and inference over private datasets, without exposing sensitive data.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    LibrePhotos

    LibrePhotos

    A self-hosted open source photo management service

    ...The system allows individuals to store and manage their photos and videos locally rather than relying on commercial cloud services. It provides features similar to services like Google Photos but runs on a private server controlled by the user. The application includes AI-powered tools that automatically analyze images to detect faces, objects, and locations, allowing photos to be grouped and searched more efficiently. LibrePhotos supports a wide variety of media formats and provides a web interface that can be accessed from different devices and operating systems. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 24

    DiffGen

    Differentially-private algorithm based on Generalization

    ...All the existing solutions that ensure epsilon-differential privacy handle the problem of disclosing relational and set-valued data in a privacy preserving manner separately. We developed an algorithm that considers both relational and set-valued data in differentially private disclosure of healthcare data.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB