Showing 16 open source projects for "llms and python-pptx"

View related business solutions
  • Build AI Apps with Gemini 3 on Vertex AI Icon
    Build AI Apps with Gemini 3 on Vertex AI

    Access Google’s most capable multimodal models. Train, test, and deploy AI with 200+ foundation models on one platform.

    Vertex AI gives developers access to Gemini 3—Google’s most advanced reasoning and coding model—plus 200+ foundation models including Claude, Llama, and Gemma. Build generative AI apps with Vertex AI Studio, customize with fine-tuning, and deploy to production with enterprise-grade MLOps. New customers get $300 in free credits.
    Try Vertex AI Free
  • Our Free Plans just got better! | Auth0 Icon
    Our Free Plans just got better! | Auth0

    With up to 25k MAUs and unlimited Okta connections, our Free Plan lets you focus on what you do best—building great apps.

    You asked, we delivered! Auth0 is excited to expand our Free and Paid plans to include more options so you can focus on building, deploying, and scaling applications without having to worry about your security. Auth0 now, thank yourself later.
    Try free now
  • 1
    Ollama Python

    Ollama Python

    Ollama Python library

    ollama-python is an open-source Python SDK that wraps the Ollama CLI, allowing seamless interaction with local large language models (LLMs) managed by Ollama. Developers use it to load models, send prompts, manage sessions, and stream responses directly from Python code. It simplifies integration of Ollama-based models into applications, supporting synchronous and streaming modes.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    Llama Stack

    Llama Stack

    Composable building blocks to build Llama Apps

    Llama-Stack is an open-source framework designed to facilitate the deployment and fine-tuning of large language models (LLMs) for various natural language processing tasks.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 3
    Prompt Declaration Language

    Prompt Declaration Language

    Prompt Declaration Language is a declarative prompt programming lang

    LLMs will continue to change the way we build software systems. They are not only useful as coding assistants, providing snipets of code, explanations, and code transformations, but they can also help replace components that could only previously be achieved with rule-based systems. Whether LLMs are used as coding assistants or software components, reliability remains an important concern. LLMs have a textual interface and the structure of useful prompts is not captured formally. Programming...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    bitnet.cpp

    bitnet.cpp

    Official inference framework for 1-bit LLMs

    bitnet.cpp is the official open-source inference framework and ecosystem designed to enable ultra-efficient execution of 1-bit large language models (LLMs), which quantize most model parameters to ternary values (-1, 0, +1) while maintaining competitive performance with full-precision counterparts. At its core is bitnet.cpp, a highly optimized C++ backend that supports fast, low-memory inference on both CPUs and GPUs, enabling models such as BitNet b1.58 to run without requiring enormous...
    Downloads: 7 This Week
    Last Update:
    See Project
  • Run Any Workload on Compute Engine VMs Icon
    Run Any Workload on Compute Engine VMs

    From dev environments to AI training, choose preset or custom VMs with 1–96 vCPUs and industry-leading 99.95% uptime SLA.

    Compute Engine delivers high-performance virtual machines for web apps, databases, containers, and AI workloads. Choose from general-purpose, compute-optimized, or GPU/TPU-accelerated machine types—or build custom VMs to match your exact specs. With live migration and automatic failover, your workloads stay online. New customers get $300 in free credits.
    Try Compute Engine
  • 5
    PydanticAI

    PydanticAI

    Agent Framework / shim to use Pydantic with LLMs

    When I first found FastAPI, I got it immediately. I was excited to find something so innovative and ergonomic built on Pydantic. Virtually every Agent Framework and LLM library in Python uses Pydantic, but when we began to use LLMs in Pydantic Logfire, I couldn't find anything that gave me the same feeling. PydanticAI is a Python Agent Framework designed to make it less painful to build production-grade applications with Generative AI. Built by the team behind Pydantic (the validation layer of the OpenAI SDK, the Anthropic SDK, LangChain, LlamaIndex, AutoGPT, Transformers, CrewAI, Instructor, and many more).
    Downloads: 1 This Week
    Last Update:
    See Project
  • 6
    Backtrack Sampler

    Backtrack Sampler

    An easy-to-understand framework for LLM samplers

    Backtrack Sampler is a framework designed for experimenting with custom sampling strategies for language models (LLMs), enabling the ability to rewind and revise generated tokens. It allows developers to create and test their own token generation strategies by providing a base structure for manipulating logits and probabilities, making it a flexible tool for those interested in fine-tuning the behavior of LLMs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    LangExtract

    LangExtract

    A Python library for extracting structured information

    LangExtract is a Python library developed by Google that leverages large language models (LLMs) to extract structured information from unstructured text—such as clinical notes, research papers, or literary works—based on user-defined instructions. It is designed to transform free-form text into reliable, schema-constrained data while maintaining traceability back to the source material.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Patchwork

    Patchwork

    Automate code reviews, patching and documentation

    Patchwork automates development gruntwork like PR reviews, bug fixing, security patching, and more using a self-hosted CLI agent and your preferred LLMs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Raglite

    Raglite

    RAGLite is a Python toolkit for Retrieval-Augmented Generation

    Raglite is a lightweight framework for building Retrieval-Augmented Generation (RAG) pipelines with minimal configuration. It connects large language models to vector databases for context-aware responses, enabling developers to prototype and deploy RAG systems quickly. Raglite focuses on simplicity and modularity for fast experimentation.
    Downloads: 0 This Week
    Last Update:
    See Project
  • AI-powered service management for IT and enterprise teams Icon
    AI-powered service management for IT and enterprise teams

    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity. Maximize operational efficiency with refreshingly simple, AI-powered Freshservice.
    Try it Free
  • 10
    Parsera

    Parsera

    Lightweight library for scraping web-sites with LLMs

    Scrape data from any website with only a link and column descriptions. Parsera is a tool designed to scrape web content, specifically handling poorly structured or messy websites.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 11
    Semantix

    Semantix

    Non-Pydantic, Non-JSON Schema, efficient AutoPrompting

    Semantix empowers developers to infuse meaning into their code through enhanced variable typing (semantic typing). By leveraging the power of large language models (LLMs) behind the scenes, Semantix transforms ordinary functions into intelligent, context-aware operations without explicit LLM calls.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Llama Cloud Services

    Llama Cloud Services

    Knowledge Agents and Management in the Cloud

    Llama Cloud Services is a suite of tools designed to facilitate the integration of large language models (LLMs) into applications. It offers components for parsing, extracting, and reporting on complex documents, streamlining the process of preparing data for LLM consumption.​
    Downloads: 1 This Week
    Last Update:
    See Project
  • 13
    bu-agent-sdk

    bu-agent-sdk

    An agent is just a for-loop

    The bu-agent-sdk from the Browser Use project is a minimalistic Python framework that defines an AI agent as a simple loop of tool calls, aiming to keep abstractions low so developers can build autonomous agents without unnecessary complexity. At its core, the agent loop repeatedly queries a large language model, interprets its output, and executes defined “tools” — functions annotated with task names — to perform actions, allowing the agent to complete tasks like arithmetic,...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Superduper

    Superduper

    Superduper: Integrate AI models and machine learning workflows

    Superduper is a Python-based framework for building end-2-end AI-data workflows and applications on your own data, integrating with major databases. It supports the latest technologies and techniques, including LLMs, vector-search, RAG, and multimodality as well as classical AI and ML paradigms. Developers may leverage Superduper by building compositional and declarative objects that out-source the details of deployment, orchestration versioning, and more to the Superduper engine. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    EasyR1

    EasyR1

    An Efficient, Scalable, Multi-Modality RL Training Framework

    EasyR1 is a streamlined training framework for building “R1-style” reasoning models from open-source LLMs with minimal boilerplate. It focuses on the full reasoning stack—data preparation, supervised fine-tuning, preference or outcome-based optimization, and lightweight evaluation—so you can iterate quickly on chain-of-thought–heavy tasks. The project’s philosophy is practicality: sensible defaults, one-command recipes, and compatibility with popular base models let you stand up experiments...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    LangChain Apps on Production with Jina

    LangChain Apps on Production with Jina

    Langchain Apps on Production with Jina & FastAPI

    Jina is an open-source framework for building scalable multi-modal AI apps on Production. LangChain is another open-source framework for building applications powered by LLMs. long-chain-serve helps you deploy your LangChain apps on Jina AI Cloud in a matter of seconds. You can benefit from the scalability and serverless architecture of the cloud without sacrificing the ease and convenience of local development. And if you prefer, you can also deploy your LangChain apps on your own...
    Downloads: 2 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB