Showing 39 open source projects for "ollama"

View related business solutions
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 1
    Qwen-Agent

    Qwen-Agent

    Agent framework and applications built upon Qwen>=3.0

    Qwen-Agent is a framework for building applications / agents using Qwen models (version 3.0+). It provides components for instruction following, tool usage (function calling), planning, memory, RAG (retrieval augmented generation), code interpreter, etc. It ships with example applications (Browser Assistant, Code Interpreter, Custom Assistant), supports GUI front-ends, backends, server setups. Agent workflow can maintain context / memory to perform multi-turn or more complex logic over time....
    Downloads: 8 This Week
    Last Update:
    See Project
  • 2
    CyberScraper 2077

    CyberScraper 2077

    A Powerful web scraper powered by LLM | OpenAI, Gemini & Ollama

    CyberScraper 2077 is not just another web scraping tool – it's a glimpse into the future of data extraction. Born from the neon-lit streets of a cyberpunk world, this AI-powered scraper uses OpenAI, Gemini and LocalLLM Models to slice through the web's defenses, extracting the data you need with unparalleled precision and style.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    OpenAI Harmony

    OpenAI Harmony

    Renderer for the harmony response format to be used with gpt-oss

    ...The format is essential for ensuring gpt-oss models operate correctly, as they are trained to rely on this structure for generating and organizing their responses. For users accessing gpt-oss through third-party providers like HuggingFace, Ollama, or vLLM, Harmony formatting is handled automatically, but developers building custom inference setups must implement it directly. With its flexible design, Harmony serves as the foundation for creating more interpretable, controlled, and extensible interactions with open-weight language models.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    PAL MCP

    PAL MCP

    The power of Claude Code / GeminiCLI / CodexCLI

    PAL MCP is an open-source Model Context Protocol (MCP) server designed to act as a powerful middleware layer that connects AI clients and tools—like Claude Code, Codex CLI, Cursor, and IDE plugins—to a broad range of underlying AI models, enabling collaborative multi-model workflows rather than relying on a single model. It lets developers orchestrate interactions across multiple models (including Gemini, OpenAI, Grok, Azure, Ollama, OpenRouter, and custom/self-hosted models), preserving conversation context seamlessly as tasks evolve and substeps run across tools. By supporting conversation threading and context passing, pal-mcp-server helps maintain continuity during complex processes like code reviews, automated planning, implementation, and validation, allowing models to “debate” or weigh in on specific subtasks for better outcomes.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • 5
    h2oGPT

    h2oGPT

    Private chat with local GPT with document, images, video, etc.

    ...The platform is designed to be secure and offline, ensuring that all data remains private and under the user's control. h2oGPT supports several AI models, including oLLaMa and Mixtral, making it a flexible tool for anyone needing advanced document analysis and AI-driven conversation in a secure, local setup.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    AI-Codereview-Gitlab

    AI-Codereview-Gitlab

    GitLab automatic code review tool based on large models

    ...The system monitors GitLab repositories and analyzes commits or merge requests using AI models to identify potential issues, coding mistakes, and quality improvements before the code is merged. By leveraging multiple large language model providers—including OpenAI, DeepSeek, ZhipuAI, or local models through Ollama—the platform allows teams to choose the AI engine that best fits their infrastructure and privacy requirements. When code changes occur, the system can automatically generate review comments and feedback that are posted directly into GitLab merge requests, allowing developers to see suggestions alongside human reviewer comments. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    GNNPCSAFT Chat

    GNNPCSAFT Chat

    Chatbot with GNNPCSAFT

    The GNNPCSAFT Chat is an implementation of our project that focuses on using Graph Neural Networks (GNN) to estimate the pure-component parameters of the Equation of State PC-SAFT. We developed this app so the scientific community can access the model's results easily. In this app, you can chat with LLM models (Gemini or Ollama) with GNNPCSAFT tools, allowing you to ask questions about the PC-SAFT parameters of various compounds, predict thermodynamic properties, and get insights into the GNNPCSAFT's performance.
    Downloads: 12 This Week
    Last Update:
    See Project
  • 8
    StudioOllamaUI

    StudioOllamaUI

    StudioOllamaUI is a local, portable interface for Ollama

    StudioOllamaUI: Portable .The easiest way to run local AI Do you want to use AI but don't know what Docker is? Does the terminal scare you? StudioOllamaUI is for you. Zero Installation: Works on a fresh Windows installation. No Python, no libraries, no drama. 100% Portable: Just like a portable browser. Unzip, run, and that's it. It doesn't clutter your registry or leave traces on your disk. AI for Everyone: No expensive GPU? No problem. Optimized to run smoothly on your CPU...
    Leader badge
    Downloads: 30 This Week
    Last Update:
    See Project
  • 9

    Spyder AI Chat Plugin

    Spyder IDE plugin providing separate chat pane for AI Assistance

    An OpenAI-compatible chat pane for Spyder 6.x. Supports OpenAI, Ollama, LM Studio, and any other server that exposes an OpenAI-compatible /v1/chat/completions endpoint. Installation with PyPi in the same environment as Spyder IDE: (spyder) $ pip install spyder-ai-chat Or from source: # clone / download and unzip the project source code, then: (spyder) $ cd spyder_ai_chat (spyder) $ pip install -e .
    Downloads: 0 This Week
    Last Update:
    See Project
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    More flexibility. More control.

    Generate interest, access liquidity without selling, and execute trades seamlessly. All in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 10
    CodinIT.dev

    CodinIT.dev

    Free, local, open-source AI app builder

    ...Deep Supabase integration means you can create UI and backend logic in one cohesive environment, while the model-agnostic architecture lets you connect to any AI, whether cloud-based (Gemini 3 Pro, GPT-5, Claude Sonnet 4.5) or local via Ollama, so you’re never locked in. All source code remains on your device and integrates seamlessly with your preferred IDE. A natural-language API enables powerful data queries and updates, automating tasks without leaving the chat interface. By running entirely locally, CodinIT.dev delivers maximum privacy, minimal latency, and smooth developer experiences free from cloud-based inconsistencies.
    Leader badge
    Downloads: 71 This Week
    Last Update:
    See Project
  • 11

    Infinite Sides

    Infinite Craft but in Pyside6 and Python with local LLM

    Infinite Craft but in Pyside6 and Python with local LLM (llama2 & others) using Ollama that also lets you create your own crafting game based on any topic Customize the game any way you like in the settings.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 12
    Shinkai: Local AI Agents

    Shinkai: Local AI Agents

    Shinkai allows you to create advanced AI (local) agents effortlessly

    Shinkai is a free, open-source AI platform that lets anyone create powerful AI agents without coding. These agents can collaborate with each other, handle complex tasks, and operate in decentralized crypto environments. Key Features: - No-Code Agent Creation - Build specialized agents (trading bots, sentiment trackers, etc.) with simple descriptions - Multi-Agent Collaboration - Agents work together to solve complex problems - Crypto Integration - Built-in support for decentralized...
    Downloads: 8 This Week
    Last Update:
    See Project
  • 13
    gpt-oss-20b

    gpt-oss-20b

    OpenAI’s compact 20B open model for fast, agentic, and local use

    ...Like its larger sibling (gpt-oss-120b), it offers adjustable reasoning depth and full chain-of-thought visibility for better interpretability. It’s released under a permissive Apache 2.0 license, allowing unrestricted commercial and research use. GPT-OSS-20B is compatible with Transformers, vLLM, Ollama, PyTorch, and other tools. It is ideal for developers building lightweight AI agents or experimenting with fine-tuning on consumer-grade hardware.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Mellum-4b-base

    Mellum-4b-base

    JetBrains’ 4B parameter code model for completions

    ...With a context window of 8,192 tokens, it excels at code completion, fill-in-the-middle tasks, and intelligent code suggestions for professional developer tools and IDEs. The model is efficient for both cloud inference with vLLM and local deployment using llama.cpp or Ollama, thanks to its bf16 precision and AMP training. While the base model is not fine-tuned for downstream tasks, it is designed to be easily adapted through supervised fine-tuning (SFT) or reinforcement learning (RL). Benchmarks on RepoBench, SAFIM, and HumanEval demonstrate its competitive performance, with specialized fine-tuned versions for Python already showing strong improvements.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB