Showing 4 open source projects for "prc-tools"

View related business solutions
  • Go from Data Warehouse to Data and AI platform with BigQuery Icon
    Go from Data Warehouse to Data and AI platform with BigQuery

    Build, train, and run ML models with simple SQL. Automate data prep, analysis, and predictions with built-in AI assistance from Gemini.

    BigQuery is more than a data warehouse—it's an autonomous data-to-AI platform. Use familiar SQL to train ML models, run time-series forecasts, and generate AI-powered insights with native Gemini integration. Built-in agents handle data engineering and data science workflows automatically. Get $300 in free credit, query 1 TB, and store 10 GB free monthly.
    Try BigQuery Free
  • Build AI Apps with Gemini 3 on Vertex AI Icon
    Build AI Apps with Gemini 3 on Vertex AI

    Access Google’s most capable multimodal models. Train, test, and deploy AI with 200+ foundation models on one platform.

    Vertex AI gives developers access to Gemini 3—Google’s most advanced reasoning and coding model—plus 200+ foundation models including Claude, Llama, and Gemma. Build generative AI apps with Vertex AI Studio, customize with fine-tuning, and deploy to production with enterprise-grade MLOps. New customers get $300 in free credits.
    Try Vertex AI Free
  • 1
    Neovim 99

    Neovim 99

    Neovim AI agent done right

    Neovim 99 is an experimental GitHub repository created by well-known developer and educator ThePrimeagen that explores what he describes as the “ideal AI workflow” for developers who want a streamlined, high-quality integration of AI tooling into real coding environments — particularly focused on tools like Neovim and agent-centric workflows. Rather than a polished end-product, this repo serves as a playground for testing, iterating, and documenting workflows that integrate AI agents directly into everyday coding tools, emphasizing rapid feedback loops, automation, and minimal friction. The project often includes configuration files, scripts, and examples that show how to coerce modern AI assistants into productive roles within editors, plugins, and terminal workflows, with a focus on “no excuses” productivity. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 2
    CodeCompanion

    CodeCompanion

    AI-powered coding, seamlessly in Neovim. Supports Anthropic, etc.

    Currently supports Anthropic, Copilot, Gemini, Ollama and OpenAI adapters.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 3
    MCP Hub

    MCP Hub

    An MCP client for Neovim that seamlessly integrates MCP servers

    mcphub.nvim is an MCP (Model Context Protocol) client plugin for Neovim that seamlessly integrates MCP servers into your editing workflow with an intuitive interface for managing, testing, and using MCP servers with your favorite chat plugins. Create your first MCP capable agent you need only 6 lines of code. Works with any langchain-supported LLM that supports tool calling (OpenAI, Anthropic, Groq, LLama etc.) Explore MCP capabilities and generate starter code with the interactive code...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 4
    KoboldAI

    KoboldAI

    Your gateway to GPT writing

    This is a browser-based front-end for AI-assisted writing with multiple local & remote AI models. It offers the standard array of tools, including Memory, Author's Note, World Info, Save & Load, adjustable AI settings, formatting options, and the ability to import existing AI Dungeon adventures. You can also turn on Adventure mode and play the game like AI Dungeon Unleashed. Stories can be played like a Novel, a text adventure game or used as a chatbot with an easy toggles to change between the multiple gameplay styles. ...
    Leader badge
    Downloads: 150 This Week
    Last Update:
    See Project
  • Easily Host LLMs and Web Apps on Cloud Run Icon
    Easily Host LLMs and Web Apps on Cloud Run

    Run everything from popular models with on-demand NVIDIA L4 GPUs to web apps without infrastructure management.

    Run frontend and backend services, batch jobs, host LLMs, and queue processing workloads without the need to manage infrastructure. Cloud Run gives you on-demand GPU access for hosting LLMs and running real-time AI—with 5-second cold starts and automatic scale-to-zero so you only pay for actual usage. New customers get $300 in free credit to start.
    Try Cloud Run Free
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB