Showing 781 open source projects for "python code"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • 1
    iX

    iX

    Autonomous GPT-4 agent platform

    IX is a platform for designing and deploying autonomous and [semi]-autonomous LLM-powered agents and workflows. IX provides a flexible and scalable solution for delegating tasks to AI-powered agents. Agents created with the platform can automate a wide variety of tasks while running in parallel and communicating with each other.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 2
    Automated Interpretability

    Automated Interpretability

    Code for Language models can explain neurons in language models paper

    The automated-interpretability repository implements tools and pipelines for automatically generating, simulating, and scoring explanations of neuron (or latent feature) behavior in neural networks. Instead of relying purely on manual, ad hoc interpretability probing, this repo aims to scale interpretability by using algorithmic methods that produce candidate explanations and assess their quality. It includes a “neuron explainer” component that, given a target neuron or latent feature,...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Qwen2.5-Coder

    Qwen2.5-Coder

    Qwen2.5-Coder is the code version of Qwen2.5, the large language model

    Qwen2.5-Coder, developed by QwenLM, is an advanced open-source code generation model designed for developers seeking powerful and diverse coding capabilities. It includes multiple model sizes—ranging from 0.5B to 32B parameters—providing solutions for a wide array of coding needs. The model supports over 92 programming languages and offers exceptional performance in generating code, debugging, and mathematical problem-solving. Qwen2.5-Coder, with its long context length of 128K tokens, is...
    Downloads: 18 This Week
    Last Update:
    See Project
  • 4
    ToRA

    ToRA

    Tool-integrated Reasoning LLM Agents

    ToRA is an open-source framework developed by Microsoft for building tool-integrated reasoning agents powered by large language models. The project focuses on improving the ability of AI systems to solve complex mathematical and analytical problems by combining natural language reasoning with external computational tools. Instead of relying solely on text generation, the system dynamically invokes tools such as symbolic solvers or programming libraries when deeper computation is required....
    Downloads: 0 This Week
    Last Update:
    See Project
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 5
    TensorFlow Privacy

    TensorFlow Privacy

    Library for training machine learning models with privacy for data

    Library for training machine learning models with privacy for training data. This repository contains the source code for TensorFlow Privacy, a Python library that includes implementations of TensorFlow optimizers for training machine learning models with differential privacy. The library comes with tutorials and analysis tools for computing the privacy guarantees provided.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Grok-1

    Grok-1

    Open-source, high-performance Mixture-of-Experts large language model

    Grok-1 is a 314-billion-parameter Mixture-of-Experts (MoE) large language model developed by xAI. Designed to optimize computational efficiency, it activates only 25% of its weights for each input token. In March 2024, xAI released Grok-1's model weights and architecture under the Apache 2.0 license, making them openly accessible to developers. The accompanying GitHub repository provides JAX example code for loading and running the model. Due to its substantial size, utilizing Grok-1...
    Leader badge
    Downloads: 31 This Week
    Last Update:
    See Project
  • 8
    Armadillo

    Armadillo

    fast C++ library for linear algebra & scientific computing

    * Fast C++ library for linear algebra (matrix maths) and scientific computing * Easy to use functions and syntax, deliberately similar to Matlab / Octave * Uses template meta-programming techniques to increase efficiency * Provides user-friendly wrappers for OpenBLAS, Intel MKL, LAPACK, ATLAS, ARPACK, SuperLU and FFTW libraries * Useful for machine learning, pattern recognition, signal processing, bioinformatics, statistics, finance, etc. * Downloads:...
    Leader badge
    Downloads: 2,758 This Week
    Last Update:
    See Project
  • 9
    GPT-2

    GPT-2

    Code for the paper Language Models are Unsupervised Multitask Learners

    This repository contains the code and model weights for GPT-2, a large-scale unsupervised language model described in the OpenAI paper “Language Models are Unsupervised Multitask Learners.” The intent is to provide a starting point for researchers and engineers to experiment with GPT-2: generate text, fine‐tune on custom datasets, explore model behavior, or study its internal phenomena. The repository includes scripts for sampling, training, downloading pre-trained models, and utilities for...
    Downloads: 12 This Week
    Last Update:
    See Project
  • Fully Managed MySQL, PostgreSQL, and SQL Server Icon
    Fully Managed MySQL, PostgreSQL, and SQL Server

    Automatic backups, patching, replication, and failover. Focus on your app, not your database.

    Cloud SQL handles your database ops end to end, so you can focus on your app.
    Try Free
  • 10
    FastChat

    FastChat

    Open platform for training, serving, and evaluating language models

    FastChat is an open platform for training, serving, and evaluating large language model-based chatbots. If you do not have enough memory, you can enable 8-bit compression by adding --load-8bit to the commands above. This can reduce memory usage by around half with slightly degraded model quality. It is compatible with the CPU, GPU, and Metal backend. Vicuna-13B with 8-bit compression can run on a single NVIDIA 3090/4080/T4/V100(16GB) GPU. In addition to that, you can add --cpu-offloading to...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    HunyuanVideo-I2V

    HunyuanVideo-I2V

    A Customizable Image-to-Video Model based on HunyuanVideo

    HunyuanVideo-I2V is a customizable image-to-video generation framework developed by Tencent, extending the capabilities of HunyuanVideo. It allows for high-quality video creation from still images, using PyTorch and providing pre-trained model weights, inference code, and customizable training options. The system includes a LoRA training code for adding special effects and enhancing video realism, aiming to offer versatile and scalable solutions for generating videos from static image inputs.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 12
    MGIE

    MGIE

    Guiding Instruction-based Image Editing via Multimodal Large Language

    MGIE—Guiding Instruction-based Image Editing—demonstrates how a multimodal LLM can parse natural-language editing instructions and then drive image transformations accordingly. The project focuses on making edits explainable and controllable: the model interprets text guidance, reasons over image content, and outputs edits aligned with user intent. It’s positioned as an ICLR 2024 Spotlight work, with code and references that show how to connect language planning to concrete image operations....
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    SPPAS

    SPPAS

    SPPAS - the automatic annotation and analyses of speech

    SPPAS is a scientific computer software package written and maintained by Brigitte Bigi of the Laboratoire Parole et Langage, in Aix-en-Provence, France. Available for free, with open source code, there is simply no other package for linguists to simple use in the automatic annotations of speech, the analyses of any kind of annotated data and the conversion of annotated files. SPPAS is able to produce automatically speech annotations from a recorded speech sound and its orthographic...
    Downloads: 20 This Week
    Last Update:
    See Project
  • 14
    sketch

    sketch

    AI code-writing assistant that understands data content

    Sketch is an open-source AI-powered data analysis assistant designed specifically for pandas users, enabling natural language interaction with tabular datasets to generate code, insights, and transformations. It works by summarizing the structure and statistical properties of a dataset and providing that context to a language model, allowing it to generate highly relevant and accurate responses tailored to the data. The tool integrates directly into pandas dataframes through an extension, making it easy to use within existing Python workflows without requiring additional IDE plugins. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    CodinIT.dev

    CodinIT.dev

    Free, local, open-source AI app builder

    CodinIT.dev is a free, local, open source AI app builder that lets you go from idea to full-stack application entirely on your machine, no coding required, just chat with AI. You can build unlimited apps with real-time previews, instant undo, and responsive, frictionless workflows. Deep Supabase integration means you can create UI and backend logic in one cohesive environment, while the model-agnostic architecture lets you connect to any AI, whether cloud-based (Gemini 3 Pro, GPT-5,...
    Leader badge
    Downloads: 62 This Week
    Last Update:
    See Project
  • 16
    Obsei

    Obsei

    Obsei is a low code AI powered automation tool

    Obsei is an automated no-code/low-code AI-powered text observation and analysis framework, designed for extracting insights from unstructured text data such as social media, reviews, and logs.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 17
    Plugins Quickstart

    Plugins Quickstart

    Get a ChatGPT plugin up and running in under 5 minutes

    plugins-quickstart is a starter project created by OpenAI to help developers build and deploy ChatGPT plugins quickly. It provides a minimal but complete example of how to structure a plugin, implement an API, and define the necessary configuration files. The repository demonstrates how a plugin can be served, authenticated, and integrated with ChatGPT for real-world use. By including both the backend code and plugin manifest, it guides developers through the end-to-end development workflow....
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    FLUX.1 Krea

    FLUX.1 Krea

    Powerful open source image generation model

    FLUX.1 Krea [dev] is an open-source 12-billion parameter image generation model developed collaboratively by Krea and Black Forest Labs, designed to deliver superior aesthetic control and high image quality. It is a rectified-flow model distilled from the original Krea 1, providing enhanced sampling efficiency through classifier-free guidance distillation. The model supports generation at resolutions between 1024 and 1280 pixels with recommended inference steps between 28 and 32 for optimal...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 19
    Shinkai: Local AI Agents

    Shinkai: Local AI Agents

    Shinkai allows you to create advanced AI (local) agents effortlessly

    Shinkai is a free, open-source AI platform that lets anyone create powerful AI agents without coding. These agents can collaborate with each other, handle complex tasks, and operate in decentralized crypto environments. Key Features: - No-Code Agent Creation - Build specialized agents (trading bots, sentiment trackers, etc.) with simple descriptions - Multi-Agent Collaboration - Agents work together to solve complex problems - Crypto Integration - Built-in support for decentralized...
    Downloads: 7 This Week
    Last Update:
    See Project
  • 20
    PromptTools

    PromptTools

    Open-source tools for prompt testing and experimentation

    Welcome to prompttools created by Hegel AI! This repo offers a set of open-source, self-hostable tools for experimenting with, testing, and evaluating LLMs, vector databases, and prompts. The core idea is to enable developers to evaluate using familiar interfaces like code, notebooks, and a local playground.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 21
    Weak-to-Strong

    Weak-to-Strong

    Implements weak-to-strong learning for training stronger ML models

    Weak-to-Strong is an OpenAI research codebase that implements the concept of weak-to-strong generalization, as described in the accompanying paper. The project provides tools for training larger “strong” models using labels or guidance generated by smaller “weak” models. Its core functionality focuses on binary classification tasks, with support for fine-tuning pretrained language models and experimenting with different loss functions, including confidence-based auxiliary losses. The...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    DeepSeek MoE

    DeepSeek MoE

    Towards Ultimate Expert Specialization in Mixture-of-Experts Language

    DeepSeek-MoE (“DeepSeek MoE”) is the DeepSeek open implementation of a Mixture-of-Experts (MoE) model architecture meant to increase parameter efficiency by activating only a subset of “expert” submodules per input. The repository introduces fine-grained expert segmentation and shared expert isolation to improve specialization while controlling compute cost. For example, their MoE variant with 16.4B parameters claims comparable or better performance to standard dense models like DeepSeek 7B...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    SSD in PyTorch 1.0

    SSD in PyTorch 1.0

    High quality, fast, modular reference implementation of SSD in PyTorch

    This repository implements SSD (Single Shot MultiBox Detector). The implementation is heavily influenced by the projects ssd.pytorch, pytorch-ssd and maskrcnn-benchmark. This repository aims to be the code base for research based on SSD. Multi-GPU training and inference: We use DistributedDataParallel, you can train or test with arbitrary GPU(s), the training schema will change accordingly. Add your own modules without pain. We abstract backbone, Detector, BoxHead, BoxPredictor, etc. You can...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    GLM-4-32B-0414

    GLM-4-32B-0414

    Open Multilingual Multimodal Chat LMs

    GLM-4-32B-0414 is a powerful open-source large language model featuring 32 billion parameters, designed to deliver performance comparable to leading models like OpenAI’s GPT series. It supports multilingual and multimodal chat capabilities with an extensive 32K token context length, making it ideal for dialogue, reasoning, and complex task completion. The model is pre-trained on 15 trillion tokens of high-quality data, including substantial synthetic reasoning datasets, and further enhanced...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Sonnet

    Sonnet

    TensorFlow-based neural network library

    Sonnet is a neural network library built on top of TensorFlow designed to provide simple, composable abstractions for machine learning research. Sonnet can be used to build neural networks for various purposes, including different types of learning. Sonnet’s programming model revolves around a single concept: modules. These modules can hold references to parameters, other modules and methods that apply some function on the user input. There are a number of predefined modules that already...
    Downloads: 4 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB