Showing 217 open source projects for "vector linux 32bit"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 1
    LangChain for Java

    LangChain for Java

    LangChain4j is an open-source Java library

    LangChain for Java is an open-source Java framework designed to simplify the development of applications powered by large language models. The library provides a unified API that allows developers to connect Java applications to multiple AI providers and embedding databases without having to implement separate integrations for each service. Its architecture includes abstractions for prompts, chat interactions, document processing, embeddings, and vector storage, enabling developers to build...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 2
    Flow Matching

    Flow Matching

    A PyTorch library for implementing flow matching algorithms

    flow_matching is a PyTorch library implementing flow matching algorithms in both continuous and discrete settings, enabling generative modeling via matching vector fields rather than diffusion. The underlying idea is to parameterize a flow (a time-dependent vector field) that transports samples from a simple base distribution to a target distribution, and train via matching of flows without requiring score estimation or noisy corruption—this can lead to more efficient or stable generative...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 3
    Moltis

    Moltis

    A Rust-native claw you can trust

    Moltis is an open-source personal AI assistant platform written in Rust that is designed to run as a fully self-hosted, local-first agent environment. It compiles the entire assistant stack, including the web interface, model routing, memory, and tools, into a single self-contained binary with no external runtime dependencies. The system supports multiple large language model providers alongside local models, enabling users to maintain privacy while still accessing cloud capabilities when...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 4
    NeMo Retriever Library

    NeMo Retriever Library

    Document content and metadata extraction microservice

    NeMo Retriever Library is a scalable microservice framework designed for extracting, structuring, and enriching content from documents to support downstream generative AI applications. It processes various document types by splitting them into components such as text, tables, charts, and images, and then applies OCR and contextual analysis to convert them into structured data formats. The system is built on NVIDIA NIM microservices, enabling high-performance parallel processing and efficient...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Our Free Plans just got better! | Auth0 Icon
    Our Free Plans just got better! | Auth0

    With up to 25k MAUs and unlimited Okta connections, our Free Plan lets you focus on what you do best—building great apps.

    You asked, we delivered! Auth0 is excited to expand our Free and Paid plans to include more options so you can focus on building, deploying, and scaling applications without having to worry about your security. Auth0 now, thank yourself later.
    Try free now
  • 5
    Self-Hosted AI Starter Kit

    Self-Hosted AI Starter Kit

    Template that quickly sets up a local AI environment

    The Self-Hosted AI Starter Kit is an open-source framework designed to help developers and teams quickly provision a local AI development environment that emphasizes control, privacy, and flexibility rather than dependence on external cloud APIs. At its core, the starter kit uses Docker Compose to orchestrate essential components like an AI workflow engine, vector database, local LLM server, and persistent storage, making it suitable for prototyping AI-driven applications without exposing...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Cherche

    Cherche

    Neural Search

    Cherche allows the creation of efficient neural search pipelines using retrievers and pre-trained language models as rankers. Cherche's main strength is its ability to build diverse and end-to-end pipelines from lexical matching, semantic matching, and collaborative filtering-based models. Cherche provides modules dedicated to summarization and question answering. These modules are compatible with Hugging Face's pre-trained models and fully integrated into neural search pipelines. Search is...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Telegram Search

    Telegram Search

    AI-powered Telegram chat backup and semantic search tool system

    Telegram Search is a self-hosted tool designed to export, back up, and intelligently search Telegram chat histories using modern AI techniques. It addresses the limitations of Telegram’s native search by enabling accurate retrieval of messages across languages through advanced tokenization and semantic understanding. Telegram Search processes chat data into searchable formats, including vector embeddings, which allow users to perform fuzzy and meaning-based searches instead of relying solely...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 8
    Jlama

    Jlama

    Jlama is a modern LLM inference engine for Java

    Jlama is a modern inference engine written entirely in Java that enables developers to run large language models locally within Java applications. Unlike frameworks that require external APIs or remote services, Jlama performs inference directly on a machine using pre-trained models. This allows organizations to integrate generative AI features into their systems while maintaining full control over data privacy and infrastructure. The engine supports a wide range of open-source model...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 9
    CAG

    CAG

    Cache-Augmented Generation: A Simple, Efficient Alternative to RAG

    CAG, or Cache-Augmented Generation, is an experimental framework that explores an alternative architecture for integrating external knowledge into large language model responses. Traditional retrieval-augmented generation systems rely on real-time retrieval of documents from databases or vector stores during inference. CAG proposes a different approach by preloading relevant knowledge into the model’s context window and precomputing the model’s key-value cache before queries are processed....
    Downloads: 1 This Week
    Last Update:
    See Project
  • Add Two Lines of Code. Get Full APM. Icon
    Add Two Lines of Code. Get Full APM.

    AppSignal installs in minutes and auto-configures dashboards, alerts, and error tracking.

    Works out of the box for Rails, Django, Express, Phoenix, and more. Monitoring exceptions and performance in no time.
    Start Free
  • 10
    Self-hosted AI Package

    Self-hosted AI Package

    Run all your local AI together in one package

    Self-hosted AI Package is an open-source Docker Compose-based starter kit that makes it easy to bootstrap a full local AI and low-code development environment with commonly used open tools, empowering developers to run LLMs and AI workflows entirely on their infrastructure. The stack typically includes Ollama for running local large language models, n8n as a low-code workflow automation platform, Supabase for database and vector storage, Open WebUI for interacting with models, Flowise for...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    AnythingLLM

    AnythingLLM

    The all-in-one Desktop & Docker AI application with full RAG and AI

    A full-stack application that enables you to turn any document, resource, or piece of content into a context that any LLM can use as references during chatting. This application allows you to pick and choose which LLM or Vector Database you want to use as well as supporting multi-user management and permissions. AnythingLLM is a full-stack application where you can use commercial off-the-shelf LLMs or popular open-source LLMs and vectorDB solutions to build a private ChatGPT with no...
    Downloads: 120 This Week
    Last Update:
    See Project
  • 12
    Aix-DB

    Aix-DB

    Based on the LangChain/LangGraph framework

    Aix-DB is an open-source intelligent data analysis platform that combines large language models with database technologies to enable conversational data exploration. The system is designed as a ChatBI solution that allows users to query datasets using natural language and receive structured insights, charts, and visualizations automatically. Built on frameworks such as LangChain and LangGraph, Aix-DB integrates retrieval-augmented generation and Text-to-SQL capabilities to convert user...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    rag-search

    rag-search

    RAG Search API

    rag-search is a lightweight Retrieval-Augmented Generation API service designed to provide structured semantic search and answer generation through a simple FastAPI backend. The project integrates web search, vector embeddings, and reranking logic to retrieve relevant context before passing it to a language model for response generation. It is built to be easily deployable, requiring only environment configuration and dependency installation to run a functional RAG service. The system...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Pathway AI Pipelines

    Pathway AI Pipelines

    Ready-to-run cloud templates for RAG

    Pathway AI Pipelines is a collection of ready-to-deploy AI pipeline templates designed to help developers rapidly build production-grade retrieval-augmented generation and enterprise search applications. The project provides end-to-end examples that connect live data sources to LLM workflows, enabling applications to stay synchronized with continuously changing information. It supports numerous connectors including local files, Google Drive, SharePoint, Kafka, PostgreSQL, and real-time APIs,...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    AI PDF Chatbot LangChain

    AI PDF Chatbot LangChain

    AI PDF chatbot agent built with LangChain & LangGraph

    AI PDF Chatbot LangChain is a full-stack template for building conversational agents that can ingest and answer questions about PDF documents. The project demonstrates how to combine LangChain and LangGraph with a vector database to enable retrieval-augmented question answering over user-provided files. It includes both frontend and backend components, making it suitable as a production starting point rather than just a minimal demo. The system parses uploaded PDFs into document chunks,...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    SimpleMem

    SimpleMem

    SimpleMem: Efficient Lifelong Memory for LLM Agents

    SimpleMem is a lightweight memory-augmented model framework that helps developers build AI applications that retain long-term context and recall relevant information without overloading model context windows. It provides easy-to-use APIs for storing structured memory entries, querying those memories using semantic search, and retrieving context to augment prompt inputs for downstream processing. Unlike monolithic systems where memory management is ad-hoc, SimpleMem formalizes a memory...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    cognee

    cognee

    Deterministic LLMs Outputs for AI Applications and AI Agents

    Cognee implements scalable, modular data pipelines that allow for creating the LLM-enriched data layer using graph and vector stores. Cognee acts a semantic memory layer, unveiling hidden connections within your data and infusing it with your company's language and principles. This self-optimizing process ensures ultra-relevant, personalized, and contextually aware LLM retrievals. Any kind of data works; unstructured text or raw media files, PDFs, tables, presentations, JSON files, and so...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    segment-geospatial

    segment-geospatial

    A Python package for segmenting geospatial data with the SAM

    The segment-geospatial package draws its inspiration from segment-anything-eo repository authored by Aliaksandr Hancharenka. To facilitate the use of the Segment Anything Model (SAM) for geospatial data, I have developed the segment-anything-py and segment-geospatial Python packages, which are now available on PyPI and conda-forge. My primary objective is to simplify the process of leveraging SAM for geospatial data analysis by enabling users to achieve this with minimal coding effort. I...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    Cheshire Cat AI

    Cheshire Cat AI

    AI agent microservice

    Cheshire Cat AI Core is an open-source framework for building customizable AI agents as scalable microservices, designed to integrate conversational intelligence into applications through an API-first architecture. It allows developers to create advanced AI assistants that can interact through WebSockets, REST APIs, and embedded chat interfaces, making it suitable for both backend services and user-facing applications. The framework includes built-in support for retrieval-augmented...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    LEANN

    LEANN

    Local RAG engine for private multimodal knowledge search on devices

    LEANN is an open source system designed to enable retrieval-augmented generation (RAG) and semantic search across personal data while running entirely on local devices. It focuses on dramatically reducing the storage overhead typically required for vector search and embedding indexes, enabling efficient large-scale knowledge retrieval on consumer hardware. LEANN introduces a storage-efficient approximate nearest neighbor index combined with on-the-fly embedding recomputation to avoid storing...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    WeKnora

    WeKnora

    LLM framework for document understanding and semantic retrieval

    WeKnora is an open source framework developed for deep document understanding and semantic information retrieval using large language models. It focuses on analyzing complex and heterogeneous documents by combining multiple processing stages such as multimodal document parsing, vector indexing, and intelligent retrieval. It follows the Retrieval-Augmented Generation (RAG) paradigm, where relevant document segments are retrieved and used by language models to generate accurate, context-aware...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    SAG

    SAG

    SQL-Driven RAG Engine

    SAG is an open-source SQL-driven retrieval-augmented generation engine that dynamically constructs knowledge graphs during query processing. Instead of relying on a static knowledge graph prepared in advance, the system automatically builds relational structures between entities while processing user queries. Documents are first decomposed into atomic semantic events, which are then represented using multidimensional natural language vectors. These vectors allow the system to identify...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    SmythOS

    SmythOS

    Cloud-native runtime for agentic AI

    SmythOS SRE (Smyth Runtime Environment) is an open-source runtime and development platform designed for building and operating production-grade AI agents. It provides a foundational infrastructure layer that functions similarly to an operating system for agentic AI systems, managing resources such as language models, storage, vector databases, and caching through a unified interface. Developers can use the runtime to create, deploy, and orchestrate intelligent agents across local machines,...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    LangChain Rust

    LangChain Rust

    LangChain for Rust, the easiest way to write LLM-based programs

    LangChain Rust is an open-source Rust implementation inspired by the LangChain ecosystem for building applications powered by large language models. The library aims to provide Rust developers with a structured framework for orchestrating prompts, chains, agents, and external tools within LLM-driven workflows. By adapting LangChain concepts to the Rust programming language, the project emphasizes performance, safety, and efficient memory management. Developers can use the framework to build...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    DeepSearcher

    DeepSearcher

    Open Source Deep Research Alternative to Reason and Search

    DeepSearcher is an open-source “deep research” style system that combines retrieval with evaluation and reasoning to answer complex questions using private or enterprise data. It is designed around the idea that high-quality answers require more than top-k retrieval, so it orchestrates multi-step search, evidence collection, and synthesis into a comprehensive response. The project integrates with vector databases (including Milvus and related options) so organizations can index internal...
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB