Open Source Python Artificial Intelligence Software - Page 73

Python Artificial Intelligence Software

View 13566 business solutions

Browse free open source Python Artificial Intelligence Software and projects below. Use the toggles on the left to filter open source Python Artificial Intelligence Software by OS, license, language, programming language, and project status.

  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 1
    Extended Dreambooth How-To Guides

    Extended Dreambooth How-To Guides

    Implementation of Dreambooth

    Extended Dreambooth How-To Guides is an implementation and extended toolkit for fine-tuning Stable Diffusion models using the DreamBooth technique, enabling users to train AI image generators to reproduce specific subjects, styles, or identities from a small set of reference images. The project adapts and expands upon earlier DreamBooth research by providing practical scripts, notebooks, and workflows that allow users to train personalized models on local machines, cloud environments, or platforms such as Google Colab. It focuses heavily on usability, offering detailed guides for different setups while still exposing advanced configuration options for experienced users. The system allows users to bind a unique token to a subject, which can later be used in prompts to generate consistent and recognizable outputs across different contexts. It also supports captioning, multi-subject training, and regularization techniques to improve generalization and avoid overfitting.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    FARM

    FARM

    Fast & easy transfer learning for NLP

    FARM makes Transfer Learning with BERT & Co simple, fast and enterprise-ready. It's built upon transformers and provides additional features to simplify the life of developers: Parallelized preprocessing, highly modular design, multi-task learning, experiment tracking, easy debugging and close integration with AWS SageMaker. With FARM you can build fast proofs-of-concept for tasks like text classification, NER or question answering and transfer them easily into production. Easy fine-tuning of language models to your task and domain language. AMP optimizers (~35% faster) and parallel preprocessing (16 CPU cores => ~16x faster). Modular design of language models and prediction heads. Switch between heads or combine them for multitask learning. Full Compatibility with HuggingFace Transformers' models and model hub. Smooth upgrading to newer language models. Integration of custom datasets via Processor class. Powerful experiment tracking & execution.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    FATE

    FATE

    An industrial grade federated learning framework

    FATE (Federated AI Technology Enabler) is the world's first industrial grade federated learning open source framework to enable enterprises and institutions to collaborate on data while protecting data security and privacy. It implements secure computation protocols based on homomorphic encryption and multi-party computation (MPC). Supporting various federated learning scenarios, FATE now provides a host of federated learning algorithms, including logistic regression, tree-based algorithms, deep learning and transfer learning. FATE became open-source in February 2019. FATE TSC was established to lead FATE open-source community, with members from major domestic cloud computing and financial service enterprises. FedAI is a community that helps businesses and organizations build AI models effectively and collaboratively, by using data in accordance with user privacy protection, data security, data confidentiality and government regulations.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    FEDML Open Source

    FEDML Open Source

    The unified and scalable ML library for large-scale training

    A Unified and Scalable Machine Learning Library for Running Training and Deployment Anywhere at Any Scale. TensorOpera AI is the next-gen cloud service for LLMs & Generative AI. It helps developers to launch complex model training, deployment, and federated learning anywhere on decentralized GPUs, multi-clouds, edge servers, and smartphones, easily, economically, and securely. Highly integrated with TensorOpera open source library, TensorOpera AI provides holistic support of three interconnected AI infrastructure layers: user-friendly MLOps, a well-managed scheduler, and high-performance ML libraries for running any AI jobs across GPU Clouds. A typical workflow is shown in the figure above. When a developer wants to run a pre-built job in Studio or Job Store, TensorOperaLaunch swiftly pairs AI jobs with the most economical GPU resources, and auto-provisions, and effortlessly runs the job, eliminating complex environment setup and management.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    More flexibility. More control.

    Generate interest, access liquidity without selling, and execute trades seamlessly. All in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 5

    FRODO 2

    Open-Source Framework for Distributed Constraint Optimization (DCOP)

    FRODO is a Java platform to solve Distributed Constraint Satisfaction Problems (DisCSPs) and Optimization Problems (DCOPs). It provides implementations for a variety of algorithms, including DPOP (and its variants), ADOPT, SynchBB, DSA...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Face Mask Detection

    Face Mask Detection

    Face Mask Detection system based on computer vision and deep learning

    Face Mask Detection system based on computer vision and deep learning using OpenCV and Tensorflow/Keras. Face Mask Detection System built with OpenCV, Keras/TensorFlow using Deep Learning and Computer Vision concepts in order to detect face masks in static images as well as in real-time video streams. Amid the ongoing COVID-19 pandemic, there are no efficient face mask detection applications which are now in high demand for transportation means, densely populated areas, residential districts, large-scale manufacturers and other enterprises to ensure safety. The absence of large datasets of ‘with_mask’ images has made this task cumbersome and challenging. Our face mask detector doesn't use any morphed masked images dataset and the model is accurate. Owing to the use of MobileNetV2 architecture, it is computationally efficient, thus making it easier to deploy the model to embedded systems (Raspberry Pi, Google Coral, etc.).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Fara-7B

    Fara-7B

    An Efficient Agentic Model for Computer Use

    Fara-7B is a Microsoft initiative aimed at bringing rigor, transparency, and structured evaluation to AI systems through automated and customizable assessment frameworks. It provides stakeholders with a way to benchmark and evaluate models across dimensions such as fairness, robustness, security, privacy, and ethical considerations. Rather than relying on ad-hoc or manual review processes, FARA enables organizations to profile AI behavior using standardized tests, metrics, and reporting templates, making evaluations reproducible and comparable over time. The framework supports plugin-based modules that can be tailored to industry-specific concerns or regulatory requirements, helping compliance teams, auditors, and engineers collaborate on shared assessment goals.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    A cross-platform library that computes fast and accurate SIFT image features. libsiftfast provides Octave/Matlab scripts, a command line interface, and a python interface (siftfastpy). Optimized with SIMD instructions and OpenMP .
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    FastChat

    FastChat

    Open platform for training, serving, and evaluating language models

    FastChat is an open platform for training, serving, and evaluating large language model-based chatbots. If you do not have enough memory, you can enable 8-bit compression by adding --load-8bit to the commands above. This can reduce memory usage by around half with slightly degraded model quality. It is compatible with the CPU, GPU, and Metal backend. Vicuna-13B with 8-bit compression can run on a single NVIDIA 3090/4080/T4/V100(16GB) GPU. In addition to that, you can add --cpu-offloading to commands above to offload weights that don't fit on your GPU onto the CPU memory. This requires 8-bit compression to be enabled and the bitsandbytes package to be installed, which is only available on linux operating systems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Fully Managed MySQL, PostgreSQL, and SQL Server Icon
    Fully Managed MySQL, PostgreSQL, and SQL Server

    Automatic backups, patching, replication, and failover. Focus on your app, not your database.

    Cloud SQL handles your database ops end to end, so you can focus on your app.
    Try Free
  • 10
    FastEdit

    FastEdit

    Editing large language models within 10 seconds

    FastEdit focuses on rapid “model editing,” letting you surgically update facts or behaviors in an LLM without full fine-tuning. It implements practical editing algorithms that insert or revise knowledge with targeted parameter updates, aiming to preserve model quality outside the edited scope. This approach is valuable when you need urgent corrections—think product names, APIs, or fast-changing facts—without retraining on large corpora. The repository provides evaluation harnesses so you can measure locality (does the change stay contained?) and generalization (does the change apply where it should?). It’s structured for repeatable experiments, making side-by-side comparisons of editing methods and hyperparameters straightforward. For applied teams, FastEdit offers a toolbox to keep models current and compliant while minimizing collateral damage to overall performance.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    FastPhotoStyle

    FastPhotoStyle

    Style transfer, deep learning, feature transform

    FastPhotoStyle is a deep learning-based image stylization framework designed to transfer the style of one photograph onto another while preserving photorealistic quality. Unlike traditional artistic style transfer methods that produce painterly outputs, this approach focuses on maintaining realistic textures, lighting, and spatial consistency. The method is based on a two-step process that includes a stylization phase followed by a smoothing operation, ensuring that the output image remains coherent and free of visual artifacts. It is computationally efficient due to its closed-form solution, allowing fast processing compared to iterative optimization-based methods. The framework is particularly useful in applications such as photo editing, film post-processing, and dataset augmentation where realism is critical. By preserving structural details and avoiding distortions, it produces results that are visually consistent with natural images.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    FastVLM

    FastVLM

    This repository contains the official implementation of FastVLM

    FastVLM is an efficiency-focused vision-language modeling stack that introduces FastViTHD, a hybrid vision encoder engineered to emit fewer visual tokens and slash encoding time, especially for high-resolution images. Instead of elaborate pruning stages, the design trades off resolution and token count through input scaling, simplifying the pipeline while maintaining strong accuracy. Reported results highlight dramatic speedups in time-to-first-token and competitive quality versus contemporary open VLMs, including comparisons across small and larger variants. The repository documents model variants, showcases head-to-head numbers against known baselines, and explains how the encoder integrates with common LLM backbones. Apple’s research brief frames FastVLM as targeting real-time or latency-sensitive scenarios, where lowering visual token pressure is critical to interactive UX. In short, it’s a practical recipe to make VLMs fast without exotic token-selection heuristics.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    FastViT

    FastViT

    This repository contains the official implementation of research

    FastViT is an efficient vision backbone family that blends convolutional inductive biases with transformer capacity to deliver strong accuracy at mobile and real-time inference budgets. Its design pursues a favorable latency-accuracy Pareto curve, targeting edge devices and server scenarios where throughput and tail latency matter. The models use lightweight attention and carefully engineered blocks to minimize token mixing costs while preserving representation power. Training and inference recipes highlight straightforward integration into common vision tasks such as classification, detection, and segmentation. The codebase provides reference implementations and checkpoints that make it easy to evaluate or fine-tune on downstream datasets. In practice, FastViT offers drop-in backbones that reduce compute and memory pressure without exotic training tricks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Feature-engine

    Feature-engine

    Feature engineering package with sklearn like functionality

    Feature-engine is a Python library with multiple transformers to engineer and select features for use in machine learning models. Feature-engine's transformers follow Scikit-learn's functionality with fit() and transform() methods to learn the transforming parameters from the data and then transform it.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Featuretools

    Featuretools

    An open source python library for automated feature engineering

    An open source Python framework for automated feature engineering. Featuretools automatically creates features from temporal and relational datasets. Featuretools uses DFS for automated feature engineering. You can combine your raw data with what you know about your data to build meaningful features for machine learning and predictive modeling. Featuretools provides APIs to ensure only valid data is used for calculations, keeping your feature vectors safe from common label leakage problems. You can specify prediction times row-by-row. Featuretools come with a library of low-level functions that can be stacked to create features. You can build and share your own custom primitives to be reused on any dataset. Featuretools works alongside tools you already use to build machine learning pipelines. You can load in pandas' data frames and automatically create meaningful features in a fraction of the time it would take to do so manually.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Feed-forward neural network for python
    ffnet is a fast and easy-to-use feed-forward neural network training solution for python. Many nice features are implemented: arbitrary network connectivity, automatic data normalization, very efficient training tools, network export to fortran code. Now ffnet has also a GUI called ffnetui.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    File-sharing-Bot

    File-sharing-Bot

    Telegram Bot to store Posts and Documents

    Telegram Bot to store posts and documents and it can be accessed by special links.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    FinGLM

    FinGLM

    Committed to building an open, public welfare

    FinGLM is an open-source financial large language model initiative aimed at advancing artificial intelligence applications within the finance industry. The project focuses on developing domain-specific language models that understand financial terminology, corporate reports, and economic datasets. By combining large language model architectures with financial datasets such as corporate annual reports and structured financial records, FinGLM aims to improve AI performance on tasks that require domain expertise. The repository also provides educational materials and tutorials that help developers learn how to build and fine-tune financial AI systems using the GLM model ecosystem. In addition to model development, the project promotes collaboration between researchers, companies, and developers interested in applying AI to financial analysis.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    FinRobot

    FinRobot

    An Open-Source AI Agent Platform for Financial Analysis using LLMs

    FinRobot is an open-source AI framework focused on automating financial data workflows by combining data ingestion, feature engineering, model training, and automated decision-making pipelines tailored for quantitative finance applications. It provides developers and quants with structured modules to fetch market data, process time series, generate technical indicators, and construct features appropriate for machine learning models, while also supporting backtesting and evaluation metrics to measure strategy performance. Built with modularity in mind, FinRobot allows users to plug in custom models — from classical algorithms to deep learning architectures — and orchestrate components in pipelines that can run reproducibly across experiments. The framework also tends to include automation layers for deployment, enabling trained models to operate in live or simulated environments with scheduled re-training and risk controls in place.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    Finance

    Finance

    150+ quantitative finance Python programs

    Finance is a repository that compiles structured notes and educational material related to financial analysis, markets, and quantitative finance concepts. The project focuses on explaining key principles used in finance and investment analysis, including topics such as financial statements, valuation models, portfolio theory, and financial markets. The repository is designed as a study reference for students and professionals who want to understand financial systems and the analytical frameworks used in financial decision-making. It organizes concepts into structured documents that explain both theoretical principles and practical calculations used in finance. The materials often include definitions, formulas, conceptual explanations, and examples to help readers understand how financial models and instruments function in real markets.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    FineSplice

    FineSplice

    Enhanced splice junction detection and estimation from RNA-Seq data

    FineSplice is a Python wrapper to TopHat2 geared towards a reliable identification of expressed exon junctions from RNA-Seq data, at enhanced detection precision with small loss in sensitivity. Following alignment with TopHat2 using known transcript annotations, FineSplice takes as input the resulting BAM file and outputs a confident set of expressed splice junctions with the corresponding read counts. Potential false positives arising from spurious alignments are filtered out via a semi-supervised anomaly detection strategy based on logistic regression. Multiple mapping reads with a unique location after filtering are rescued and reallocated to the most reliable candidate location. FineSplice requires Python 2.x (>= 2.6) with the following modules installed: pysam (http://code.google.com/p/pysam/) and scikit-learn (http://scikit-learn.org/). For further details check out our publication: Nucl. Acids Res. (2014) doi: 10.1093/nar/gku166
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    FireRedASR

    FireRedASR

    Open-source industrial-grade ASR models

    FireRedASR is an industrial-grade family of open-source automatic speech recognition models designed to provide high-precision speech-to-text performance across languages including Mandarin, English, and various Chinese dialects, achieving new state-of-the-art benchmarks on public test sets. The project includes multiple model variants to meet different application needs, such as high-accuracy end-to-end interaction using an encoder-adapter-LLM framework and efficient real-time recognition using attention-based encoder-decoder architectures, giving developers flexibility in balancing performance and resource constraints. FireRedASR not only excels in traditional speech recognition tasks but also demonstrates strong capability in challenging scenarios like singing lyrics recognition, where accurate transcription is often difficult for conventional models.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    FireRedTTS-2

    FireRedTTS-2

    Long-form streaming TTS system for multi-speaker dialogue generation

    FireRedTTS2 is a next-generation open-source text-to-speech (TTS) system focused on long-form, streaming speech synthesis for multi-speaker dialogue, delivering stable natural speech with context-aware prosody and reliable speaker transitions that support real-time and conversational applications. It features a specialized streaming speech tokenizer and a dual-transformer architecture that enables low latency and high-quality synthesis, making it suitable for interactive systems like chatbots, podcasts, and applications where dynamic turn-taking between speakers is essential. FireRedTTS2 supports multilingual output and speaker flexibility, enabling scenarios that involve language switching, cross-lingual voice cloning, and expressive dialogue generation that maintains consistency over longer utterances.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    Firefly LLM

    Firefly LLM

    A large model training tool that supports training large models

    Firefly is an open-source framework designed to simplify the training and fine-tuning of large language models through a unified and configurable workflow. The project provides a comprehensive environment where developers can perform tasks such as model pre-training, instruction tuning, and preference optimization using widely adopted machine learning techniques. Its architecture supports both full-parameter training and parameter-efficient strategies like LoRA and QLoRA, making it suitable for environments with limited computational resources. Firefly is compatible with a wide range of popular open-source models including LLaMA, Qwen, Baichuan, InternLM, and Mistral, enabling developers to experiment with different architectures using a consistent training pipeline. The framework also provides curated datasets and training templates that help streamline the process of instruction tuning and conversational model development.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Five video classification methods

    Five video classification methods

    Code that accompanies my blog post outlining five video classification

    Classifying video presents unique challenges for machine learning models. As I’ve covered in my previous posts, video has the added (and interesting) property of temporal features in addition to the spatial features present in 2D images. While this additional information provides us more to work with, it also requires different network architectures and, often, adds larger memory and computational demands.We won’t use any optical flow images. This reduces model complexity, training time, and a whole whack load of hyperparameters we don’t have to worry about. Every video will be subsampled down to 40 frames. So a 41-frame video and a 500-frame video will both be reduced to 40 frames, with the 500-frame video essentially being fast-forwarded. We won’t do much preprocessing. A common preprocessing step for video classification is subtracting the mean, but we’ll keep the frames pretty raw from start to finish.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB