28 Integrations with Database Mart

View a list of Database Mart integrations and software that integrates with Database Mart below. Compare the best Database Mart integrations as well as features, ratings, user reviews, and pricing of software that integrates with Database Mart. Here are the current Database Mart integrations in 2026:

  • 1
    TensorFlow

    TensorFlow

    TensorFlow

    An end-to-end open source machine learning platform. TensorFlow is an end-to-end open source platform for machine learning. It has a comprehensive, flexible ecosystem of tools, libraries and community resources that lets researchers push the state-of-the-art in ML and developers easily build and deploy ML powered applications. Build and train ML models easily using intuitive high-level APIs like Keras with eager execution, which makes for immediate model iteration and easy debugging. Easily train and deploy models in the cloud, on-prem, in the browser, or on-device no matter what language you use. A simple and flexible architecture to take new ideas from concept to code, to state-of-the-art models, and to publication faster. Build, deploy, and experiment easily with TensorFlow.
    Starting Price: Free
  • 2
    Kubernetes

    Kubernetes

    Kubernetes

    Kubernetes (K8s) is an open-source system for automating deployment, scaling, and management of containerized applications. It groups containers that make up an application into logical units for easy management and discovery. Kubernetes builds upon 15 years of experience of running production workloads at Google, combined with best-of-breed ideas and practices from the community. Designed on the same principles that allows Google to run billions of containers a week, Kubernetes can scale without increasing your ops team. Whether testing locally or running a global enterprise, Kubernetes flexibility grows with you to deliver your applications consistently and easily no matter how complex your need is. Kubernetes is open source giving you the freedom to take advantage of on-premises, hybrid, or public cloud infrastructure, letting you effortlessly move workloads to where it matters to you.
    Starting Price: Free
  • 3
    MySQL

    MySQL

    Oracle

    MySQL is the world's most popular open source database. With its proven performance, reliability, and ease-of-use, MySQL has become the leading database choice for web-based applications, used by high profile web properties including Facebook, Twitter, YouTube, and all five of the top five websites*. Additionally, it is an extremely popular choice as embedded database, distributed by thousands of ISVs and OEMs.
    Starting Price: Free
  • 4
    SQL Server

    SQL Server

    Microsoft

    Intelligence and security are built into Microsoft SQL Server 2019. You get extras without extra cost, along with best-in-class performance and flexibility for your on-premises needs. Take advantage of the efficiency and agility of the cloud by easily migrating to the cloud without changing code. Unlock insights and make predictions faster with Azure. Develop using the technology of your choice, including open source, backed by Microsoft's innovations. Easily integrate data into your apps and use a rich set of cognitive services to build human-like intelligence across any scale of data. AI is native to the data platform—you can unlock insights faster from all your data, on-premises and in the cloud. Combine your unique enterprise data and the world's data to build an intelligence-driven organization. Work with a flexible data platform that gives you a consistent experience across platforms and gets your innovations to market faster—you can build your apps and then deploy anywhere.
    Starting Price: Free
  • 5
    Mistral AI

    Mistral AI

    Mistral AI

    Mistral AI is a pioneering artificial intelligence startup specializing in open-source generative AI. The company offers a range of customizable, enterprise-grade AI solutions deployable across various platforms, including on-premises, cloud, edge, and devices. Flagship products include "Le Chat," a multilingual AI assistant designed to enhance productivity in both personal and professional contexts, and "La Plateforme," a developer platform that enables the creation and deployment of AI-powered applications. Committed to transparency and innovation, Mistral AI positions itself as a leading independent AI lab, contributing significantly to open-source AI and policy development.
    Starting Price: Free
  • 6
    Redis

    Redis

    Redis Labs

    Redis Labs: home of Redis. Redis Enterprise is the best version of Redis. Go beyond cache; try Redis Enterprise free in the cloud using NoSQL & data caching with the world’s fastest in-memory database. Run Redis at scale, enterprise grade resiliency, massive scalability, ease of management, and operational simplicity. DevOps love Redis in the Cloud. Developers can access enhanced data structures, a variety of modules, and rapid innovation with faster time to market. CIOs love the confidence of working with 99.999% uptime best in class security and expert support from the creators of Redis. Implement relational databases, active-active, geo-distribution, built in conflict distribution for simple and complex data types, & reads/writes in multiple geo regions to the same data set. Redis Enterprise offers flexible deployment options, cloud on-prem, & hybrid. Redis Labs: home of Redis. Redis JSON, Redis Java, Python Redis, Redis on Kubernetes & Redis gui best practices.
    Starting Price: Free
  • 7
    Apache Cassandra

    Apache Cassandra

    Apache Software Foundation

    The Apache Cassandra database is the right choice when you need scalability and high availability without compromising performance. Linear scalability and proven fault-tolerance on commodity hardware or cloud infrastructure make it the perfect platform for mission-critical data. Cassandra's support for replicating across multiple datacenters is best-in-class, providing lower latency for your users and the peace of mind of knowing that you can survive regional outages.
  • 8
    DeepSeek R1

    DeepSeek R1

    DeepSeek

    DeepSeek-R1 is an advanced open-source reasoning model developed by DeepSeek, designed to rival OpenAI's Model o1. Accessible via web, app, and API, it excels in complex tasks such as mathematics and coding, demonstrating superior performance on benchmarks like the American Invitational Mathematics Examination (AIME) and MATH. DeepSeek-R1 employs a mixture of experts (MoE) architecture with 671 billion total parameters, activating 37 billion parameters per token, enabling efficient and accurate reasoning capabilities. This model is part of DeepSeek's commitment to advancing artificial general intelligence (AGI) through open-source innovation.
    Starting Price: Free
  • 9
    Keras

    Keras

    Keras

    Keras is an API designed for human beings, not machines. Keras follows best practices for reducing cognitive load: it offers consistent & simple APIs, it minimizes the number of user actions required for common use cases, and it provides clear & actionable error messages. It also has extensive documentation and developer guides. Keras is the most used deep learning framework among top-5 winning teams on Kaggle. Because Keras makes it easier to run new experiments, it empowers you to try more ideas than your competition, faster. And this is how you win. Built on top of TensorFlow 2.0, Keras is an industry-strength framework that can scale to large clusters of GPUs or an entire TPU pod. It's not only possible; it's easy. Take advantage of the full deployment capabilities of the TensorFlow platform. You can export Keras models to JavaScript to run directly in the browser, to TF Lite to run on iOS, Android, and embedded devices. It's also easy to serve Keras models as via a web API.
  • 10
    LangChain

    LangChain

    LangChain

    LangChain is a powerful, composable framework designed for building, running, and managing applications powered by large language models (LLMs). It offers an array of tools for creating context-aware, reasoning applications, allowing businesses to leverage their own data and APIs to enhance functionality. LangChain’s suite includes LangGraph for orchestrating agent-driven workflows, and LangSmith for agent observability and performance management. Whether you're building prototypes or scaling full applications, LangChain offers the flexibility and tools needed to optimize the LLM lifecycle, with seamless integrations and fault-tolerant scalability.
  • 11
    Qwen2.5

    Qwen2.5

    Alibaba

    Qwen2.5 is an advanced multimodal AI model designed to provide highly accurate and context-aware responses across a wide range of applications. It builds on the capabilities of its predecessors, integrating cutting-edge natural language understanding with enhanced reasoning, creativity, and multimodal processing. Qwen2.5 can seamlessly analyze and generate text, interpret images, and interact with complex data to deliver precise solutions in real time. Optimized for adaptability, it excels in personalized assistance, data analysis, creative content generation, and academic research, making it a versatile tool for professionals and everyday users alike. Its user-centric design emphasizes transparency, efficiency, and alignment with ethical AI practices.
    Starting Price: Free
  • 12
    Hugging Face

    Hugging Face

    Hugging Face

    Hugging Face is a leading platform for AI and machine learning, offering a vast hub for models, datasets, and tools for natural language processing (NLP) and beyond. The platform supports a wide range of applications, from text, image, and audio to 3D data analysis. Hugging Face fosters collaboration among researchers, developers, and companies by providing open-source tools like Transformers, Diffusers, and Tokenizers. It enables users to build, share, and access pre-trained models, accelerating AI development for a variety of industries.
    Starting Price: $9 per month
  • 13
    Milvus

    Milvus

    Zilliz

    Vector database built for scalable similarity search. Open-source, highly scalable, and blazing fast. Store, index, and manage massive embedding vectors generated by deep neural networks and other machine learning (ML) models. With Milvus vector database, you can create a large-scale similarity search service in less than a minute. Simple and intuitive SDKs are also available for a variety of different languages. Milvus is hardware efficient and provides advanced indexing algorithms, achieving a 10x performance boost in retrieval speed. Milvus vector database has been battle-tested by over a thousand enterprise users in a variety of use cases. With extensive isolation of individual system components, Milvus is highly resilient and reliable. The distributed and high-throughput nature of Milvus makes it a natural fit for serving large-scale vector data. Milvus vector database adopts a systemic approach to cloud-nativity, separating compute from storage.
    Starting Price: Free
  • 14
    Chroma

    Chroma

    Chroma

    Chroma is an AI-native open-source embedding database. Chroma has all the tools you need to use embeddings. Chroma is building the database that learns. Pick up an issue, create a PR, or participate in our Discord and let the community know what features you would like.
    Starting Price: Free
  • 15
    Ollama

    Ollama

    Ollama

    Ollama is an innovative platform that focuses on providing AI-powered tools and services, designed to make it easier for users to interact with and build AI-driven applications. Run AI models locally. By offering a range of solutions, including natural language processing models and customizable AI features, Ollama empowers developers, businesses, and organizations to integrate advanced machine learning technologies into their workflows. With an emphasis on usability and accessibility, Ollama strives to simplify the process of working with AI, making it an appealing option for those looking to harness the potential of artificial intelligence in their projects.
    Starting Price: Free
  • 16
    Llama 3.1
    The open source AI model you can fine-tune, distill and deploy anywhere. Our latest instruction-tuned model is available in 8B, 70B and 405B versions. Using our open ecosystem, build faster with a selection of differentiated product offerings to support your use cases. Choose from real-time inference or batch inference services. Download model weights to further optimize cost per token. Adapt for your application, improve with synthetic data and deploy on-prem or in the cloud. Use Llama system components and extend the model using zero shot tool use and RAG to build agentic behaviors. Leverage 405B high quality data to improve specialized models for specific use cases.
    Starting Price: Free
  • 17
    ComfyUI

    ComfyUI

    ComfyUI

    ComfyUI is a free and open source node-based application for generative AI, enabling users to build, create, and share without limits. It allows for the extension of functionality through custom nodes, letting users tailor workflows to their specific needs. Designed for performance, ComfyUI runs workflows directly on local machines, offering faster iteration, lower costs, and complete control. The visual interface provides full control by connecting nodes on a canvas, allowing for branching, remixing, and adjusting every part of the workflow at any time. Workflows can be saved, shared, and reused effortlessly, with exported media carrying metadata to instantly rebuild the full workflow. Users can see results in real-time as they adjust workflows, facilitating faster iteration with instant visual feedback. ComfyUI supports the generation of various media types, including images, videos, 3D assets, and audio.
    Starting Price: Free
  • 18
    Stable Diffusion

    Stable Diffusion

    Stability AI

    Over the last few weeks we all have been overwhelmed by the response and have been working hard to ensure a safe and ethical release, incorporating data from our beta model tests and community for the developers to act on. In cooperation with the tireless legal, ethics and technology teams at HuggingFace and amazing engineers at CoreWeave. We have developed an AI-based Safety Classifier included by default in the overall software package. This understands concepts and other factors in generations to remove outputs that may not be desired by the model user. The parameters of this can be readily adjusted and we welcome input from the community how to improve this. Image generation models are powerful, but still need to improve to understand how to represent what we want better.
    Starting Price: $0.2 per image
  • 19
    PostgreSQL

    PostgreSQL

    PostgreSQL Global Development Group

    PostgreSQL is a powerful, open-source object-relational database system with over 30 years of active development that has earned it a strong reputation for reliability, feature robustness, and performance. There is a wealth of information to be found describing how to install and use PostgreSQL through the official documentation. The open-source community provides many helpful places to become familiar with PostgreSQL, discover how it works, and find career opportunities. Learm more on how to engage with the community. The PostgreSQL Global Development Group has released an update to all supported versions of PostgreSQL, including 15.1, 14.6, 13.9, 12.13, 11.18, and 10.23. This release fixes 25 bugs reported over the last several months. This is the final release of PostgreSQL 10. PostgreSQL 10 will no longer receive security and bug fixes. If you are running PostgreSQL 10 in a production environment, we suggest that you make plans to upgrade.
  • 20
    PyTorch

    PyTorch

    PyTorch

    Transition seamlessly between eager and graph modes with TorchScript, and accelerate the path to production with TorchServe. Scalable distributed training and performance optimization in research and production is enabled by the torch-distributed backend. A rich ecosystem of tools and libraries extends PyTorch and supports development in computer vision, NLP and more. PyTorch is well supported on major cloud platforms, providing frictionless development and easy scaling. Select your preferences and run the install command. Stable represents the most currently tested and supported version of PyTorch. This should be suitable for many users. Preview is available if you want the latest, not fully tested and supported, 1.10 builds that are generated nightly. Please ensure that you have met the prerequisites (e.g., numpy), depending on your package manager. Anaconda is our recommended package manager since it installs all dependencies.
  • 21
    Qdrant

    Qdrant

    Qdrant

    Qdrant is a vector similarity engine & vector database. It deploys as an API service providing search for the nearest high-dimensional vectors. With Qdrant, embeddings or neural network encoders can be turned into full-fledged applications for matching, searching, recommending, and much more! Provides the OpenAPI v3 specification to generate a client library in almost any programming language. Alternatively utilise ready-made client for Python or other programming languages with additional functionality. Implement a unique custom modification of the HNSW algorithm for Approximate Nearest Neighbor Search. Search with a State-of-the-Art speed and apply search filters without compromising on results. Support additional payload associated with vectors. Not only stores payload but also allows filter results based on payload values.
  • 22
    Phi-2

    Phi-2

    Microsoft

    We are now releasing Phi-2, a 2.7 billion-parameter language model that demonstrates outstanding reasoning and language understanding capabilities, showcasing state-of-the-art performance among base language models with less than 13 billion parameters. On complex benchmarks Phi-2 matches or outperforms models up to 25x larger, thanks to new innovations in model scaling and training data curation. With its compact size, Phi-2 is an ideal playground for researchers, including for exploration around mechanistic interpretability, safety improvements, or fine-tuning experimentation on a variety of tasks. We have made Phi-2 available in the Azure AI Studio model catalog to foster research and development on language models.
  • 23
    Gemma 2

    Gemma 2

    Google

    A family of state-of-the-art, light-open models created from the same research and technology that were used to create Gemini models. These models incorporate comprehensive security measures and help ensure responsible and reliable AI solutions through selected data sets and rigorous adjustments. Gemma models achieve exceptional comparative results in their 2B, 7B, 9B, and 27B sizes, even outperforming some larger open models. With Keras 3.0, enjoy seamless compatibility with JAX, TensorFlow, and PyTorch, allowing you to effortlessly choose and change frameworks based on task. Redesigned to deliver outstanding performance and unmatched efficiency, Gemma 2 is optimized for incredibly fast inference on various hardware. The Gemma family of models offers different models that are optimized for specific use cases and adapt to your needs. Gemma models are large text-to-text lightweight language models with a decoder, trained in a huge set of text data, code, and mathematical content.
  • 24
    Phi-3

    Phi-3

    Microsoft

    A family of powerful, small language models (SLMs) with groundbreaking performance at low cost and low latency. Maximize AI capabilities, lower resource use, and ensure cost-effective generative AI deployments across your applications. Accelerate response times in real-time interactions, autonomous systems, apps requiring low latency, and other critical scenarios. Run Phi-3 in the cloud, at the edge, or on device, resulting in greater deployment and operation flexibility. Phi-3 models were developed in accordance with Microsoft AI principles: accountability, transparency, fairness, reliability and safety, privacy and security, and inclusiveness. Operate effectively in offline environments where data privacy is paramount or connectivity is limited. Generate more coherent, accurate, and contextually relevant outputs with an expanded context window. Deploy at the edge to deliver faster responses.
  • 25
    Phi-4

    Phi-4

    Microsoft

    Phi-4 is a 14B parameter state-of-the-art small language model (SLM) that excels at complex reasoning in areas such as math, in addition to conventional language processing. Phi-4 is the latest member of our Phi family of small language models and demonstrates what’s possible as we continue to probe the boundaries of SLMs. Phi-4 is currently available on Azure AI Foundry under a Microsoft Research License Agreement (MSRLA) and will be available on Hugging Face. Phi-4 outperforms comparable and larger models on math related reasoning due to advancements throughout the processes, including the use of high-quality synthetic datasets, curation of high-quality organic data, and post-training innovations. Phi-4 continues to push the frontier of size vs quality.
  • 26
    vLLM

    vLLM

    vLLM

    vLLM is a high-performance library designed to facilitate efficient inference and serving of Large Language Models (LLMs). Originally developed in the Sky Computing Lab at UC Berkeley, vLLM has evolved into a community-driven project with contributions from both academia and industry. It offers state-of-the-art serving throughput by efficiently managing attention key and value memory through its PagedAttention mechanism. It supports continuous batching of incoming requests and utilizes optimized CUDA kernels, including integration with FlashAttention and FlashInfer, to enhance model execution speed. Additionally, vLLM provides quantization support for GPTQ, AWQ, INT4, INT8, and FP8, as well as speculative decoding capabilities. Users benefit from seamless integration with popular Hugging Face models, support for various decoding algorithms such as parallel sampling and beam search, and compatibility with NVIDIA GPUs, AMD CPUs and GPUs, Intel CPUs, and more.
  • 27
    MariaDB

    MariaDB

    MariaDB

    MariaDB Platform is a complete enterprise open source database solution. It has the versatility to support transactional, analytical and hybrid workloads as well as relational, JSON and hybrid data models. And it has the scalability to grow from standalone databases and data warehouses to fully distributed SQL for executing millions of transactions per second and performing interactive, ad hoc analytics on billions of rows. MariaDB can be deployed on prem on commodity hardware, is available on all major public clouds and through MariaDB SkySQL as a fully managed cloud database. To learn more, visit mariadb.com.
  • 28
    Oracle Database
    Oracle database products offer customers cost-optimized and high-performance versions of Oracle Database, the world's leading converged, multi-model database management system, as well as in-memory, NoSQL, and MySQL databases. Oracle Autonomous Database, available on-premises via Oracle Cloud@Customer or in the Oracle Cloud Infrastructure, enables customers to simplify relational database environments and reduce management workloads. Oracle Autonomous Database eliminates the complexity of operating and securing Oracle Database while giving customers the highest levels of performance, scalability, and availability. Oracle Database can be deployed on-premises when customers have data residency and network latency concerns. Customers with applications that are dependent on specific Oracle database versions have complete control over the versions they run and when those versions change.
  • Previous
  • You're on page 1
  • Next