Open Source Python Artificial Intelligence Software - Page 66

Python Artificial Intelligence Software

View 13566 business solutions

Browse free open source Python Artificial Intelligence Software and projects below. Use the toggles on the left to filter open source Python Artificial Intelligence Software by OS, license, language, programming language, and project status.

  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 1
    CakeChat

    CakeChat

    CakeChat: Emotional Generative Dialog System

    CakeChat is a backend for chatbots that are able to express emotions via conversations. The code is flexible and allows to condition model's responses by an arbitrary categorical variable. For example, you can train your own persona-based neural conversational model or create an emotional chatting machine. Hierarchical Recurrent Encoder-Decoder (HRED) architecture for handling deep dialog context. Multilayer RNN with GRU cells. The first layer of the utterance-level encoder is always bidirectional. By default, CuDNNGRU implementation is used for ~25% acceleration during inference. Thought vector is fed into decoder on each decoding step. Decoder can be conditioned on any categorical label, for example, emotion label or persona id. May be initialized using w2v model trained on your corpus. Embedding layer may be either fixed or fine-tuned along with other weights of the network.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    CapsGNN

    CapsGNN

    A PyTorch implementation of "Capsule Graph Neural Network"

    A PyTorch implementation of "Capsule Graph Neural Network" (ICLR 2019). The high-quality node embeddings learned from the Graph Neural Networks (GNNs) have been applied to a wide range of node-based applications and some of them have achieved state-of-the-art (SOTA) performance. However, when applying node embeddings learned from GNNs to generate graph embeddings, the scalar node representation may not suffice to preserve the node/graph properties efficiently, resulting in sub-optimal graph embeddings. Inspired by the Capsule Neural Network (CapsNet), we propose the Capsule Graph Neural Network (CapsGNN), which adopts the concept of capsules to address the weakness in existing GNN-based graph embeddings algorithms. By extracting node features in the form of capsules, routing mechanism can be utilized to capture important information at the graph level. As a result, our model generates multiple embeddings for each graph to capture graph properties from different aspects.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Cathnet is developing the infrastructure for the Catholic Semantic Web. Technologies involved include, but are not limited to, XML, RDF, NLP, Zope, Plone and Plone products.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Cellicone is a project to develop an artificial life organism with the necessary components to make it comparable to biological life as we know it. This includes components ranging from proteins to cells to organs to limbs, and many steps between.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 5
    Provides a voice interface for applications via a plug in system. Allows the inclusion of voice recognition in an application with a minimum of effort.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    ChainerCV

    ChainerCV

    ChainerCV: a Library for Deep Learning in Computer Vision

    ChainerCV is a collection of tools to train and run neural networks for computer vision tasks using Chainer. In ChainerCV, we define the object detection task as a problem of, given an image, bounding box-based localization and categorization of objects. Bounding boxes in an image are represented as a two-dimensional array of shape (R,4), where R is the number of bounding boxes and the second axis corresponds to the coordinates of bounding boxes. ChainerCV supports dataset loaders, which can be used to easily index examples with list-like interfaces. Dataset classes whose names end with BboxDataset contain annotations of where objects locate in an image and which categories they are assigned to. These datasets can be indexed to return a tuple of an image, bounding boxes and labels. ChainerCV provides several network implementations that carry out object detection.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    ChainerRL

    ChainerRL

    ChainerRL is a deep reinforcement learning library

    ChainerRL (this repository) is a deep reinforcement learning library that implements various state-of-the-art deep reinforcement algorithms in Python using Chainer, a flexible deep learning framework. PFRL is the PyTorch analog of ChainerRL. ChainerRL has a set of accompanying visualization tools in order to aid developers' ability to understand and debug their RL agents. With this visualization tool, the behavior of ChainerRL agents can be easily inspected from a browser UI. Environments that support the subset of OpenAI Gym's interface (reset and step methods) can be used.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Chameleon LLM

    Chameleon LLM

    Codes for "Chameleon: Plug-and-Play Compositional Reasoning

    Discover Chameleon, our cutting-edge compositional reasoning framework designed to enhance large language models (LLMs) and overcome their inherent limitations, such as outdated information and lack of precise reasoning. By integrating various tools such as vision models, web search engines, Python functions, and rule-based modules, Chameleon delivers more accurate, up-to-date, and precise responses, making it a game-changer in the natural language processing landscape. With GPT-4 at its core, Chameleon has showcased exceptional improvements in accuracy on benchmark tasks, outperforming competitors and setting new industry standards.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Charlemagne is a versatile genetic programming application which includes a command-line client and an interactive console mode. It features built in input-output mapping support, and is user-extensible for complex fitness evaluation in Python and Lisp
    Downloads: 0 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 10
    ChatGLM Efficient Tuning

    ChatGLM Efficient Tuning

    Fine-tuning ChatGLM-6B with PEFT

    ChatGLM-Efficient-Tuning is a hands-on toolkit for fine-tuning ChatGLM-family models with parameter-efficient methods on everyday hardware. It wraps techniques like LoRA and prompt-tuning into simple training scripts so you can adapt a large model to your domain without full retraining. The project exposes practical switches for quantization and mixed precision, allowing bigger models to fit into limited VRAM. It includes examples for instruction tuning and dialogue datasets, making it straightforward to stand up a task-specific assistant. Because the code leans on widely used libraries, you can bring your own datasets and monitoring tools with minimal glue. For builders who want results fast, it’s a pragmatic way to specialize ChatGLM while controlling costs and turnaround time.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    ChatGPT API

    ChatGPT API

    Reverse engineered ChatGPT API

    acheong08/ChatGPT is a community-built project that packaged an unofficial, reverse-engineered interface for interacting with ChatGPT, primarily aimed at developers who wanted programmatic access before official solutions were widely available. It focuses on enabling chatbot-like workflows by wrapping the underlying web-based behavior into a developer-friendly library and tooling, making it easier to integrate into scripts, bots, and experiments. The repository is structured like an open-source client project, including installation guidance, documentation, and contribution processes, and it emphasizes extensibility for different use cases. Because it is not an official product, it is best understood as an experimental bridge that tried to translate web interactions into a reusable API surface. Projects like this often prioritize rapid iteration and practicality, while requiring users to be mindful of stability and changes in upstream services.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    ChatGPT Academic

    ChatGPT Academic

    ChatGPT extension for scientific research work

    ChatGPT extension for scientific research work, specially optimized academic paper polishing experience, supports custom shortcut buttons, supports custom function plug-ins, supports markdown table display, double display of Tex formulas, complete code display function, new local Python/C++/Go project tree Analysis function/Project source code self-translation ability, newly added PDF and Word document batch summary function/PDF paper full-text translation function. All buttons are dynamically generated by reading functional.py, you can add custom functions at will, and liberate the pasteboard. Support for markdown tables output by GPT. If the output contains a formula, it will be displayed in tex form and rendered form at the same time, which is convenient for copying and reading.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    ChatGPT Discord Bot

    ChatGPT Discord Bot

    Integrate ChatGPT into your own discord bot

    Build your own Discord bot using ChatGPT. Using certain personas may generate vulgar or disturbing content. Use at your own risk. Public mode (default), the bot directly replies on the channel. The bot's reply can only be seen by the person who used the command. Invite your bot to your server via OAuth2 URL Generator. Email/Password authentication (Not supported for Google/Microsoft accounts). A system prompt would be invoked when the bot is first started or reset. You can set it up by modifying the content in system_prompt.txt. All the text in the file will be fired as a prompt to the bot. Get the first message from ChatGPT in your discord channel!
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    ChatGPT-Reviewer

    ChatGPT-Reviewer

    Automated pull requests reviewing and issues triaging with ChatGPT

    Automated pull requests reviewing and issues triaging with ChatGPT. Create an OpenAI API key here, and then set the key as an action secret in your repository named OPENAI_API_KEY. The ChatGPT reviewer PRs are also getting reviewed by ChatGPT, refer the pull requests for the sample review comments. In order to protect public repositories for malicious users, Github runs all pull request workflows raised from repository forks with a read-only token and no access to secrets.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    ChatGenTitle

    ChatGenTitle

    A paper title generation model fine-tuned on the LLaMA model

    ChatGenTitle: A paper title generation model fine-tuned on the LLaMA model using information from millions of arXiv papers.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Chat_Bot

    Chat_Bot

    A chat bot that talks to users on Omegle Site

    Here is a chat bot that will talk to users on Omegle Site, it uses AIML, python and requires connection to internet to work. You'll be amused to see the bot chat with other people on Omegle.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Cheshire Cat AI

    Cheshire Cat AI

    AI agent microservice

    Cheshire Cat AI Core is an open-source framework for building customizable AI agents as scalable microservices, designed to integrate conversational intelligence into applications through an API-first architecture. It allows developers to create advanced AI assistants that can interact through WebSockets, REST APIs, and embedded chat interfaces, making it suitable for both backend services and user-facing applications. The framework includes built-in support for retrieval-augmented generation using vector databases such as Qdrant, enabling agents to incorporate external knowledge and documents into their responses. It is highly extensible through a plugin system that supports custom tools, event hooks, and workflows, giving developers fine-grained control over agent behavior and interactions. Cheshire Cat also supports multi-user environments with granular permissions and identity provider integration, making it suitable for enterprise use cases.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    Chinese Llama 2 7B

    Chinese Llama 2 7B

    The first Chinese LLaMA2 model in the open source community

    Chinese Llama 2 7B is an open-source large language model adapted from the LLaMA-2 architecture and optimized for Chinese and bilingual Chinese-English applications. The project provides a version of LLaMA-2 that has been further trained on Chinese data so it can better understand and generate text in Chinese while maintaining compatibility with the original model ecosystem. In addition to the model weights, the repository also includes supervised fine-tuning datasets and training resources that help developers build chat-optimized versions of the model. The project follows the input format used by the LLaMA-2 chat architecture, ensuring compatibility with existing optimization techniques and tools built for the LLaMA-2 ecosystem. By releasing both the model and associated datasets, the project allows researchers and developers to experiment with Chinese language models in a fully open environment.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    Chinese-LLaMA-Alpaca 2

    Chinese-LLaMA-Alpaca 2

    Chinese LLaMA-2 & Alpaca-2 Large Model Phase II Project

    This project is developed based on the commercially available large model Llama-2 released by Meta. It is the second phase of the Chinese LLaMA&Alpaca large model project. The Chinese LLaMA-2 base model and the Alpaca-2 instruction fine-tuning large model are open-sourced. These models expand and optimize the Chinese vocabulary on the basis of the original Llama-2, use large-scale Chinese data for incremental pre-training, and further improve the basic semantics and command understanding of Chinese. Performance improvements. The related model supports FlashAttention-2 training, supports 4K context and can be extended up to 18K+ through the NTK method.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    Chinese-LLaMA-Alpaca-2 v2.0

    Chinese-LLaMA-Alpaca-2 v2.0

    Chinese LLaMA & Alpaca large language model + local CPU/GPU training

    This project has open-sourced the Chinese LLaMA model and the Alpaca large model with instruction fine-tuning to further promote the open research of large models in the Chinese NLP community. Based on the original LLaMA , these models expand the Chinese vocabulary and use Chinese data for secondary pre-training, which further improves the basic semantic understanding of Chinese. At the same time, the Chinese Alpaca model further uses Chinese instruction data for fine-tuning, which significantly improves the model's ability to understand and execute instructions.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    Chinese-LLaMA-Alpaca-3

    Chinese-LLaMA-Alpaca-3

    Chinese Llama-3 LLMs) developed from Meta Llama 3

    Chinese-LLaMA-Alpaca-3 is an open-source project that provides Mandarin-focused large language models based on Meta’s LLaMA-3 architecture, with both foundational and instruction-tuned variants to support high-quality Chinese natural language understanding and generation. It extends the original LLaMA models with expanded Chinese vocabularies and additional pretraining on Chinese corpora to improve semantic encoding and decoding specifically for Chinese text. Alongside the base models, the project also releases Chinese Alpaca models that are fine-tuned on instruction datasets so they behave more like conversational and instruction-following AI assistants. It includes scripts and tooling that let researchers or developers run training, fine-tuning, quantization, and deployment on local machines (CPU or GPU), making experimentation and testing accessible without requiring large clusters.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    Chinese-XLNet

    Chinese-XLNet

    Chinese XLNet pre-trained model

    Chinese-XLNet is a Chinese language pre-trained model based on the XLNet architecture, providing an advanced foundation for natural language processing tasks in Mandarin and other Chinese dialects. Unlike traditional masked language modeling, XLNet uses a permutation language modeling objective that captures bidirectional context more effectively by training over all possible token orderings, yielding richer contextual representations. This model is trained on large-scale Chinese text datasets to learn linguistic patterns, long-range dependencies, and semantic nuance typical of Chinese writing, making it useful for tasks like text classification, question answering, named entity recognition, and language generation. Chinese-XLNet offers an alternative to models like BERT by emphasizing autoregressive and permutation-based learning, which can lead to performance improvements on certain benchmarks and tasks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Chloe - a chatterbot with charisma. Learns from experience, and speaks any language you teach her. Perhaps even useful, but surely entertaining.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24

    Chronological Cohesive Units

    The experimental source code for the paper

    The experimental source code for the paper, "A Novel Recommendation Approach Based on Chronological Cohesive Units in Content Consuming"
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Chronos Forecasting

    Chronos Forecasting

    Pretrained (Language) Models for Probabilistic Time Series Forecasting

    Chronos is a family of pretrained time series forecasting models based on language model architectures. A time series is transformed into a sequence of tokens via scaling and quantization, and a language model is trained on these tokens using the cross-entropy loss. Once trained, probabilistic forecasts are obtained by sampling multiple future trajectories given the historical context. Chronos models have been trained on a large corpus of publicly available time series data, as well as synthetic data generated using Gaussian processes.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB