Business Software for Langflow - Page 2

Top Software that integrates with Langflow as of June 2025 - Page 2

Langflow Clear Filters
  • 1
    Mistral Small

    Mistral Small

    Mistral AI

    On September 17, 2024, Mistral AI announced several key updates to enhance the accessibility and performance of their AI offerings. They introduced a free tier on "La Plateforme," their serverless platform for tuning and deploying Mistral models as API endpoints, enabling developers to experiment and prototype at no cost. Additionally, Mistral AI reduced prices across their entire model lineup, with significant cuts such as a 50% reduction for Mistral Nemo and an 80% decrease for Mistral Small and Codestral, making advanced AI more cost-effective for users. The company also unveiled Mistral Small v24.09, a 22-billion-parameter model offering a balance between performance and efficiency, suitable for tasks like translation, summarization, and sentiment analysis. Furthermore, they made Pixtral 12B, a vision-capable model with image understanding capabilities, freely available on "Le Chat," allowing users to analyze and caption images without compromising text-based performance.
    Starting Price: Free
  • 2
    Upstash

    Upstash

    Upstash

    Blend the performance of in-memory with the durability of disk storage enabling many use cases beyond caching. You can leverage multi-region replication with global databases. True Serverless Kafka where price scales to zero. With per-request pricing, you only pay what you use. Built-in REST API allows you to produce and consume your Kafka topics from anywhere. Start free, then pay only for what you use with per-request pricing. Forget your expensive server/instance. Use Upstash as much as you need, you'll never pay more than the cap price, guaranteed. Upstash REST API enables access from Cloudflare Workers and Fastly Compute@Edge. With the global database, you can access your database from anywhere with very low latency. Low latency data, ease of use, and pay-per-request pricing makes Upstash a perfect choice for the Jamstack and Serverless world. With servers/instances, you pay per hour or a fixed price. With Serverless, you pay per request.
    Starting Price: $0.2 per 100K commands
  • 3
    Airbyte

    Airbyte

    Airbyte

    Airbyte is an open-source data integration platform designed to help businesses synchronize data from various sources to their data warehouses, lakes, or databases. The platform provides over 550 pre-built connectors and enables users to easily create custom connectors using low-code or no-code tools. Airbyte's solution is optimized for large-scale data movement, enhancing AI workflows by seamlessly integrating unstructured data into vector databases like Pinecone and Weaviate. It offers flexible deployment options, ensuring security, compliance, and governance across all models.
    Starting Price: $2.50 per credit
  • 4
    Hugging Face

    Hugging Face

    Hugging Face

    Hugging Face is a leading platform for AI and machine learning, offering a vast hub for models, datasets, and tools for natural language processing (NLP) and beyond. The platform supports a wide range of applications, from text, image, and audio to 3D data analysis. Hugging Face fosters collaboration among researchers, developers, and companies by providing open-source tools like Transformers, Diffusers, and Tokenizers. It enables users to build, share, and access pre-trained models, accelerating AI development for a variety of industries.
    Starting Price: $9 per month
  • 5
    Chroma

    Chroma

    Chroma

    Chroma is an AI-native open-source embedding database. Chroma has all the tools you need to use embeddings. Chroma is building the database that learns. Pick up an issue, create a PR, or participate in our Discord and let the community know what features you would like.
    Starting Price: Free
  • 6
    Ollama

    Ollama

    Ollama

    Ollama is an innovative platform that focuses on providing AI-powered tools and services, designed to make it easier for users to interact with and build AI-driven applications. Run AI models locally. By offering a range of solutions, including natural language processing models and customizable AI features, Ollama empowers developers, businesses, and organizations to integrate advanced machine learning technologies into their workflows. With an emphasis on usability and accessibility, Ollama strives to simplify the process of working with AI, making it an appealing option for those looking to harness the potential of artificial intelligence in their projects.
    Starting Price: Free
  • 7
    Mixtral 8x7B

    Mixtral 8x7B

    Mistral AI

    Mixtral 8x7B is a high-quality sparse mixture of experts model (SMoE) with open weights. Licensed under Apache 2.0. Mixtral outperforms Llama 2 70B on most benchmarks with 6x faster inference. It is the strongest open-weight model with a permissive license and the best model overall regarding cost/performance trade-offs. In particular, it matches or outperforms GPT-3.5 on most standard benchmarks.
    Starting Price: Free
  • 8
    Zep

    Zep

    Zep

    Zep ensures your assistant remembers past conversations and resurfaces them when relevant. Identify your user's intent, build semantic routers, and trigger events, all in milliseconds. Emails, phone numbers, dates, names, and more, are extracted quickly and accurately. Your assistant will never forget a user. Classify intent, emotion, and more and turn dialog into structured data. Retrieve, analyze, and extract in milliseconds; your users never wait. We don't send your data to third-party LLM services. SDKs for your favorite languages and frameworks. Automagically populate prompts with a summary of relevant past conversations, no matter how distant. Zep summarizes, embeds, and executes retrieval pipelines over your Assistant's chat history. Instantly and accurately classify chat dialog. Understand user intent and emotion. Route chains based on semantic context, and trigger events. Quickly extract business data from chat conversations.
    Starting Price: Free
  • 9
    Codestral

    Codestral

    Mistral AI

    We introduce Codestral, our first-ever code model. Codestral is an open-weight generative AI model explicitly designed for code generation tasks. It helps developers write and interact with code through a shared instruction and completion API endpoint. As it masters code and English, it can be used to design advanced AI applications for software developers. Codestral is trained on a diverse dataset of 80+ programming languages, including the most popular ones, such as Python, Java, C, C++, JavaScript, and Bash. It also performs well on more specific ones like Swift and Fortran. This broad language base ensures Codestral can assist developers in various coding environments and projects.
    Starting Price: Free
  • 10
    Meta AI
    Meta AI is an intelligent assistant that is capable of complex reasoning, following instructions, visualizing ideas, and solving nuanced problems. Meta AI is an intelligent assistant built on Meta's most advanced model. It is designed to answer any question you might have, help with writing, provide step-by-step advice, and create images to share with friends. It is available within Meta's family of apps, smart glasses, and web platforms.
    Starting Price: Free
  • 11
    Mistral Large

    Mistral Large

    Mistral AI

    Mistral Large is Mistral AI's flagship language model, designed for advanced text generation and complex multilingual reasoning tasks, including text comprehension, transformation, and code generation. It supports English, French, Spanish, German, and Italian, offering a nuanced understanding of grammar and cultural contexts. With a 32,000-token context window, it can accurately recall information from extensive documents. The model's precise instruction-following and native function-calling capabilities facilitate application development and tech stack modernization. Mistral Large is accessible through Mistral's platform, Azure AI Studio, and Azure Machine Learning, and can be self-deployed for sensitive use cases. Benchmark evaluations indicate that Mistral Large achieves strong results, making it the world's second-ranked model generally available through an API, next to GPT-4.
    Starting Price: Free
  • 12
    Instructor

    Instructor

    Instructor

    Instructor is a tool that enables developers to extract structured data from natural language using Large Language Models (LLMs). Integrating with Python's Pydantic library allows users to define desired output structures through type hints, facilitating schema validation and seamless integration with IDEs. Instructor supports various LLM providers, including OpenAI, Anthropic, Litellm, and Cohere, offering flexibility in implementation. Its customizable nature permits the definition of validators and custom error messages, enhancing data validation processes. Instructor is trusted by engineers from platforms like Langflow, underscoring its reliability and effectiveness in managing structured outputs powered by LLMs. Instructor is powered by Pydantic, which is powered by type hints. Schema validation and prompting are controlled by type annotations; less to learn, and less code to write, and it integrates with your IDE.
    Starting Price: Free
  • 13
    Milvus

    Milvus

    Milvus

    We are an intelligent HelpDesk, we will optimize management processes, increase the productivity of your team and increase the efficiency of your support. Offer more possibilities to your customers, opening tickets from different devices. Have more flexibility to configure your customers' SLA in a unique, simple and direct way. Use the mobile app of the inventory management system and ensure the good management of IT assets, through the monitoring and total control of the entire equipment park of your customers wherever and whenever you want. Increase the productivity of your technical support team with intelligent and automated inventory control , see some of the resources. Optimize your management and relationship with your customers! Get alerts of key machine features on your dashboard. Workflow, automation and ticket triggers. Password Vault, satisfaction survey, ticket scheduling, follow customer, service catalog, follow ticket, ticket conference, advanced dashboard.
    Starting Price: $25 per month
  • 14
    Yahoo Finance

    Yahoo Finance

    Yahoo Finance

    The planet’s biggest business news platform. Sharing business news you need, plus live shows every weekday. Yahoo Finance provides free stock quotes, up-to-date news, portfolio management resources, international market data, charting, social interaction and mortgage rates that help you manage your financial life.
  • 15
    Pinecone

    Pinecone

    Pinecone

    The AI Knowledge Platform. The Pinecone Database, Inference, and Assistant make building high-performance vector search apps easy. Developer-friendly, fully managed, and easily scalable without infrastructure hassles. Once you have vector embeddings, manage and search through them in Pinecone to power semantic search, recommenders, and other applications that rely on relevant information retrieval. Ultra-low query latency, even with billions of items. Give users a great experience. Live index updates when you add, edit, or delete data. Your data is ready right away. Combine vector search with metadata filters for more relevant and faster results. Launch, use, and scale your vector search service with our easy API, without worrying about infrastructure or algorithms. We'll keep it running smoothly and securely.
  • 16
    Vectara

    Vectara

    Vectara

    Vectara is LLM-powered search-as-a-service. The platform provides a complete ML search pipeline from extraction and indexing to retrieval, re-ranking and calibration. Every element of the platform is API-addressable. Developers can embed the most advanced NLP models for app and site search in minutes. Vectara automatically extracts text from PDF and Office to JSON, HTML, XML, CommonMark, and many more. Encode at scale with cutting edge zero-shot models using deep neural networks optimized for language understanding. Segment data into any number of indexes storing vector encodings optimized for low latency and high recall. Recall candidate results from millions of documents using cutting-edge, zero-shot neural network models. Increase the precision of retrieved results with cross-attentional neural networks to merge and reorder results. Zero in on the true likelihoods that the retrieved response represents a probable answer to the query.
    Starting Price: Free
  • 17
    Pixtral Large

    Pixtral Large

    Mistral AI

    Pixtral Large is a 124-billion-parameter open-weight multimodal model developed by Mistral AI, building upon their Mistral Large 2 architecture. It integrates a 123-billion-parameter multimodal decoder with a 1-billion-parameter vision encoder, enabling advanced understanding of documents, charts, and natural images while maintaining leading text comprehension capabilities. With a context window of 128,000 tokens, Pixtral Large can process at least 30 high-resolution images simultaneously. The model has demonstrated state-of-the-art performance on benchmarks such as MathVista, DocVQA, and VQAv2, surpassing models like GPT-4o and Gemini-1.5 Pro. Pixtral Large is available under the Mistral Research License for research and educational use, and under the Mistral Commercial License for commercial applications.
    Starting Price: Free
  • 18
    Composio

    Composio

    Composio

    Composio is an integration platform designed to enhance AI agents and Large Language Models (LLMs) by providing seamless connections to over 150 tools with minimal code. It supports a wide array of agentic frameworks and LLM providers, facilitating function calling for efficient task execution. Composio offers a comprehensive repository of tools, including GitHub, Salesforce, file management systems, and code execution environments, enabling AI agents to perform diverse actions and subscribe to various triggers. The platform features managed authentication, allowing users to oversee authentication processes for all users and agents from a centralized dashboard. Composio's core capabilities include a developer-first integration approach, built-in authentication management, an expanding catalog of over 90 ready-to-connect tools, a 30% increase in reliability through simplified JSON structures and improved error handling, SOC Type II compliance ensuring maximum data security.
    Starting Price: $49 per month
  • 19
    AI Crypto-Kit
    AI Crypto-Kit empowers developers to build crypto agents by seamlessly integrating leading Web3 platforms like Coinbase, OpenSea, and more to automate real-world crypto/DeFi workflows. Developers can build AI-powered crypto automation in minutes, including applications such as trading agents, community reward systems, Coinbase wallet management, portfolio tracking, market analysis, and yield farming. The platform offers capabilities engineered for crypto agents, including fully managed agent authentication with support for OAuth, API keys, JWT, and automatic token refresh; optimization for LLM function calling to ensure enterprise-grade reliability; support for over 20 agentic frameworks like Pippin, LangChain, and LlamaIndex; integration with more than 30 Web3 platforms, including Binance, Aave, OpenSea, and Chainlink; and SDKs and APIs for agentic app interactions, available in Python and TypeScript.
  • 20
    Couchbase

    Couchbase

    Couchbase

    Unlike other NoSQL databases, Couchbase provides an enterprise-class, multicloud to edge database that offers the robust capabilities required for business-critical applications on a highly scalable and available platform. As a distributed cloud-native database, Couchbase runs in modern dynamic environments and on any cloud, either customer-managed or fully managed as-a-service. Couchbase is built on open standards, combining the best of NoSQL with the power and familiarity of SQL, to simplify the transition from mainframe and relational databases. Couchbase Server is a multipurpose, distributed database that fuses the strengths of relational databases such as SQL and ACID transactions with JSON’s versatility, with a foundation that is extremely fast and scalable. It’s used across industries for things like user profiles, dynamic product catalogs, GenAI apps, vector search, high-speed caching, and much more.
  • 21
    Glean

    Glean

    Glean Technologies

    Founded by former Google search engineers, Glean understands context, language, behavior, and relationships with others, to find personalized answers to your questions— instantly. Glean learns your company’s unique language and continuously trains to improve search performance. Glean reveals insights you never knew existed, and makes connections with the people who can help. So everyone is on the same page and can focus where they need to. Quick set-up. Instant performance. Get started in less than 2 hours. Glean searches across your company's collective knowledge and into your content. No need to remember where things are or what they’re called. Search for and understand who people are, what they’re working on, and how they can help.
  • 22
    Qdrant

    Qdrant

    Qdrant

    Qdrant is a vector similarity engine & vector database. It deploys as an API service providing search for the nearest high-dimensional vectors. With Qdrant, embeddings or neural network encoders can be turned into full-fledged applications for matching, searching, recommending, and much more! Provides the OpenAPI v3 specification to generate a client library in almost any programming language. Alternatively utilise ready-made client for Python or other programming languages with additional functionality. Implement a unique custom modification of the HNSW algorithm for Approximate Nearest Neighbor Search. Search with a State-of-the-Art speed and apply search filters without compromising on results. Support additional payload associated with vectors. Not only stores payload but also allows filter results based on payload values.
  • 23
    Amazon Bedrock
    Amazon Bedrock is a fully managed service that simplifies building and scaling generative AI applications by providing access to a variety of high-performing foundation models (FMs) from leading AI companies such as AI21 Labs, Anthropic, Cohere, Meta, Mistral AI, Stability AI, and Amazon itself. Through a single API, developers can experiment with these models, customize them using techniques like fine-tuning and Retrieval Augmented Generation (RAG), and create agents that interact with enterprise systems and data sources. As a serverless platform, Amazon Bedrock eliminates the need for infrastructure management, allowing seamless integration of generative AI capabilities into applications with a focus on security, privacy, and responsible AI practices.
  • 24
    Wolfram Alpha

    Wolfram Alpha

    Wolfram Alpha

    The introduction of Wolfram Alpha defined a fundamentally new paradigm for getting knowledge and answers—not by searching the web, but by doing dynamic computations based on a vast collection of built-in data, algorithms and methods. Wolfram Alpha brings expert-level knowledge and capabilities to the broadest possible range of people—spanning all professions and education levels. We work to accept completely free-form input, and to serve as a knowledge engine that generates powerful results and presents them with maximum clarity. What makes Wolfram Alpha possible today is a somewhat unique set of circumstances—and the singular vision of Stephen Wolfram. For the first time in history, computers are powerful enough to support the capabilities of Wolfram Alpha, and the web provides a broad-based means of delivery. But this technology alone was not enough for Wolfram Alpha to be possible.
  • 25
    Unstructured

    Unstructured

    Unstructured

    80% of enterprise data exists in difficult-to-use formats like HTML, PDF, CSV, PNG, PPTX, and more. Unstructured effortlessly extracts and transforms complex data for use with every major vector database and LLM framework. Unstructured allows data scientists to pre-process data at scale so they spend less time collecting and cleaning, and more time modeling and analyzing. Our enterprise-grade connectors capture data wherever it lives, so we can transform it into AI-friendly JSON files for companies who are eager to fold AI into their business. You can count on Unstructured to deliver data that's curated, clean of artifacts, and most importantly, LLM-ready.
  • 26
    Groq

    Groq

    Groq

    Groq is on a mission to set the standard for GenAI inference speed, helping real-time AI applications come to life today. An LPU inference engine, with LPU standing for Language Processing Unit, is a new type of end-to-end processing unit system that provides the fastest inference for computationally intensive applications with a sequential component, such as AI language applications (LLMs). The LPU is designed to overcome the two LLM bottlenecks, compute density and memory bandwidth. An LPU has greater computing capacity than a GPU and CPU in regards to LLMs. This reduces the amount of time per word calculated, allowing sequences of text to be generated much faster. Additionally, eliminating external memory bottlenecks enables the LPU inference engine to deliver orders of magnitude better performance on LLMs compared to GPUs. Groq supports standard machine learning frameworks such as PyTorch, TensorFlow, and ONNX for inference.
  • 27
    Le Chat

    Le Chat

    Mistral AI

    Le Chat is a conversational entry point to interact with the various models from Mistral AI. It offers a pedagogical and fun way to explore Mistral AI’s technology. Le Chat can use Mistral Large or Mistral Small under the hood, or a prototype model called Mistral Next, designed to be brief and concise. We are hard at work to make our models as useful and as little opinionated as possible, although much remain to be improved! Thanks to a tunable system-level moderation mechanism, Le Chat warns you in a non-invasive way when you’re pushing the conversation in directions where the assistant may produce sensitive or controversial content.
    Starting Price: Free
  • 28
    Cake AI

    Cake AI

    Cake AI

    Cake AI is a comprehensive AI infrastructure platform that enables teams to build and deploy AI applications using hundreds of pre-integrated open source components, offering complete visibility and control. It provides a curated, end-to-end selection of fully managed, best-in-class commercial and open source AI tools, with pre-built integrations across the full breadth of components needed to move an AI application into production. Cake supports dynamic autoscaling, comprehensive security measures including role-based access control and encryption, advanced monitoring, and infrastructure flexibility across various environments, including Kubernetes clusters and cloud services such as AWS. Its data layer equips teams with tools for data ingestion, transformation, and analytics, leveraging tools like Airflow, DBT, Prefect, Metabase, and Superset. For AI operations, Cake integrates with model catalogs like Hugging Face and supports modular workflows using LangChain, LlamaIndex, and more.
  • 29
    DataStax

    DataStax

    DataStax

    The Open, Multi-Cloud Stack for Modern Data Apps. Built on open-source Apache Cassandra™. Global-scale and 100% uptime without vendor lock-in. Deploy on multi-cloud, on-prem, open-source, and Kubernetes. Elastic and pay-as-you-go for improved TCO. Start building faster with Stargate APIs for NoSQL, real-time, reactive, JSON, REST, and GraphQL. Skip the complexity of multiple OSS projects and APIs that don’t scale. Ideal for commerce, mobile, AI/ML, IoT, microservices, social, gaming, and richly interactive applications that must scale-up and scale-down with demand. Get building modern data applications with Astra, a database-as-a-service powered by Apache Cassandra™. Use REST, GraphQL, JSON with your favorite full-stack framework Richly interactive apps that are elastic and viral-ready from Day 1. Pay-as-you-go Apache Cassandra DBaaS that scales effortlessly and affordably.
  • 30
    NVIDIA DRIVE
    Software is what turns a vehicle into an intelligent machine. The NVIDIA DRIVE™ Software stack is open, empowering developers to efficiently build and deploy a variety of state-of-the-art AV applications, including perception, localization and mapping, planning and control, driver monitoring, and natural language processing. The foundation of the DRIVE Software stack, DRIVE OS is the first safe operating system for accelerated computing. It includes NvMedia for sensor input processing, NVIDIA CUDA® libraries for efficient parallel computing implementations, NVIDIA TensorRT™ for real-time AI inference, and other developer tools and modules to access hardware engines. The NVIDIA DriveWorks® SDK provides middleware functions on top of DRIVE OS that are fundamental to autonomous vehicle development. These consist of the sensor abstraction layer (SAL) and sensor plugins, data recorder, vehicle I/O support, and a deep neural network (DNN) framework.