Showing 75 open source projects for "meta-extractor"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Fully Managed MySQL, PostgreSQL, and SQL Server Icon
    Fully Managed MySQL, PostgreSQL, and SQL Server

    Automatic backups, patching, replication, and failover. Focus on your app, not your database.

    Cloud SQL handles your database ops end to end, so you can focus on your app.
    Try Free
  • 1
    Meta-World

    Meta-World

    Collections of robotics environments

    Meta-World is an open-source benchmark suite of robotic manipulation environments focused on multi-task and meta reinforcement learning. It provides a large collection of continuous-control tasks, such as reaching, pushing, opening doors, and manipulating objects with a simulated robot arm. The library defines standardized benchmarks like MT1, MT10, and MT50 for multi-task learning, where a single policy is trained across different numbers of tasks.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 2
    Meta Agents Research Environments (ARE)

    Meta Agents Research Environments (ARE)

    Meta Agents Research Environments is a comprehensive platform

    Meta Agents Research Environments (ARE) is a simulation and benchmarking platform. It is designed to evaluate AI agents in dynamic, evolving, multi-step tasks. Unlike static benchmarks, ARE supports environments where agents must adapt to changes over time and reason over sequences of actions. It interacts with applications and faces uncertainty.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Llama Recipes

    Llama Recipes

    Scripts for fine-tuning Meta Llama3 with composable FSDP & PEFT method

    The 'llama-recipes' repository is a companion to the Meta Llama models. We support the latest version, Llama 3.1, in this repository. The goal is to provide a scalable library for fine-tuning Meta Llama models, along with some example scripts and notebooks to quickly get started with using the models in a variety of use-cases, including fine-tuning for domain adaptation and building LLM-based applications with Llama and other tools in the LLM ecosystem.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    MLJ.jl

    MLJ.jl

    A Julia machine learning framework

    MLJ (Machine Learning in Julia) is a toolbox written in Julia providing a common interface and meta-algorithms for selecting, tuning, evaluating, composing, and comparing about 200 machine learning models written in Julia and other languages. The functionality of MLJ is distributed over several repositories illustrated in the dependency chart below. These repositories live at the JuliaAI umbrella organization.
    Downloads: 1 This Week
    Last Update:
    See Project
  • Train ML Models With SQL You Already Know Icon
    Train ML Models With SQL You Already Know

    BigQuery automates data prep, analysis, and predictions with built-in AI assistance.

    Build and deploy ML models using familiar SQL. Automate data prep with built-in Gemini. Query 1 TB and store 10 GB free monthly.
    Try Free
  • 5
    Get Shit Done

    Get Shit Done

    A light-weight and powerful meta-prompting, context engineering

    Get Shit Done is a high-impact, open-source meta-prompting and spec-driven development system designed to streamline building software with AI assistants like Claude Code, OpenCode, and Gemini CLI. It solves “context rot” — the degradation of AI quality as a chat session grows — by structuring your idea into precise, context-engineered steps that are researched, scoped, planned, executed, and verified with clear commands and outputs instead of ad-hoc prompts.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 6
    LLaMA 3

    LLaMA 3

    The official Meta Llama 3 GitHub site

    ...It introduced the public packaging of weights, licenses, and quickstart examples that helped developers fine-tune or run the models locally and on common serving stacks. As the Llama stack evolved, Meta consolidated repositories and marked this one deprecated, pointing users to newer, centralized hubs for models, utilities, and docs. Even as a deprecated repo, it documents the transition path and preserves references that clarify how Llama 3 releases map into the current ecosystem. Practically, it functioned as a bridge between Llama 2 and later Llama releases by standardizing distribution and starter code for inference and fine-tuning. ...
    Downloads: 9 This Week
    Last Update:
    See Project
  • 7
    GPT4All

    GPT4All

    Run Local LLMs on Any Device. Open-source

    GPT4All is an open-source project that allows users to run large language models (LLMs) locally on their desktops or laptops, eliminating the need for API calls or GPUs. The software provides a simple, user-friendly application that can be downloaded and run on various platforms, including Windows, macOS, and Ubuntu, without requiring specialized hardware. It integrates with the llama.cpp implementation and supports multiple LLMs, allowing users to interact with AI models privately. This...
    Downloads: 131 This Week
    Last Update:
    See Project
  • 8
    MetaClaw

    MetaClaw

    Just talk to your agent

    MetaClaw is an AI or agent-oriented system that appears to focus on advanced control, coordination, or training of autonomous agents, potentially within reinforcement learning or tool-using environments. The project likely emphasizes meta-level reasoning, where agents are not only executing tasks but also adapting their strategies based on feedback and performance signals. It may incorporate mechanisms for learning from interactions, improving decision-making over time, and generalizing across different domains. The architecture suggests scalability, allowing the system to handle multiple agents or complex workflows simultaneously. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    llama.cpp Python Bindings

    llama.cpp Python Bindings

    Python bindings for llama.cpp

    llama-cpp-python provides Python bindings for llama.cpp, enabling the integration of LLaMA (Large Language Model Meta AI) language models into Python applications. This facilitates the use of LLaMA's capabilities in natural language processing tasks within Python environments.
    Downloads: 9 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 10
    openclaw-kapso-whatsapp

    openclaw-kapso-whatsapp

    Give your OpenClaw AI agent a WhatsApp number

    openclaw-kapso-whatsapp is a plugin repository designed to extend the OpenClaw AI agent by giving it a dedicated WhatsApp phone number using the official Meta Cloud API via Kapso, enabling direct interaction through one of the most widely used messaging platforms. This integration allows the autonomous AI assistant to send and receive messages on WhatsApp, turning the agent into a real-world task performer accessible through text conversations. The plugin is built in Go and handles communication entirely through cloud APIs, avoiding the risk of bans that come with unofficial or reverse-engineered interfaces. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 11
    TRIBE v2

    TRIBE v2

    A multimodal model for brain response prediction

    TRIBE v2 is a multimodal foundation model developed by Meta AI for predicting human brain activity from naturalistic stimuli such as video, audio, and text. It is designed for in-silico neuroscience, enabling researchers to model how the brain responds to complex real-world inputs. The system integrates state-of-the-art encoders—including LLaMA for text, V-JEPA for video, and Wav2Vec-BERT for audio—into a unified Transformer architecture.
    Downloads: 18 This Week
    Last Update:
    See Project
  • 12
    Voice-Pro

    Voice-Pro

    Comprehensive Gradio WebUI for audio processing

    Voice-Pro is the best gradio WebUI for transcription, translation and text-to-speech. It can be easily installed with one click. Create a virtual environment using Miniconda, running completely separate from the Windows system (fully portable). Supports real-time transcription and translation, as well as batch mode.
    Downloads: 18 This Week
    Last Update:
    See Project
  • 13
    MetaCLIP

    MetaCLIP

    ICLR2024 Spotlight: curation/training code, metadata, distribution

    MetaCLIP is a research codebase that extends the CLIP framework into a meta-learning / continual learning regime, aiming to adapt CLIP-style models to new tasks or domains efficiently. The goal is to preserve CLIP’s strong zero-shot transfer capability while enabling fast adaptation to domain shifts or novel class sets with minimal data and without catastrophic forgetting. The repository provides training logic, adaptation strategies (e.g. prompt tuning, adapter modules), and evaluation across base and target domains to measure how well the model retains its general knowledge while specializing as needed. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Poetiq

    Poetiq

    Reproduction of Poetiq's record-breaking submission to the ARC-AGI-1

    ...The repository allows others to reproduce their results, experiment with different LLM backends (e.g. the user may supply keys for supported models), and observe how their adaptive meta-system handles the logic and abstraction challenges.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Jittor

    Jittor

    Jittor is a high-performance deep learning framework

    Jittor is a high-performance deep learning framework based on JIT compiling and meta-operators. The whole framework and meta-operators are compiled just in time. A powerful op compiler and tuner are integrated into Jittor. It allowed us to generate high-performance code specialized for your model. Jittor also contains a wealth of high-performance model libraries, including image recognition, detection, segmentation, generation, differentiable rendering, geometric learning, reinforcement learning, etc. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Qlib

    Qlib

    Qlib is an AI-oriented quantitative investment platform

    Qlib is an AI-oriented quantitative investment platform, which aims to realize the potential, empower the research, and create the value of AI technologies in quantitative investment. With Qlib, you can easily try your ideas to create better Quant investment strategies. An increasing number of SOTA Quant research works/papers are released in Qlib. With Qlib, users can easily try their ideas to create better Quant investment strategies. At the module level, Qlib is a platform that consists of...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 17
    GSD 2

    GSD 2

    A powerful meta-prompting, context engineering

    GSD 2 is a project focused on automating and streamlining development workflows through structured build systems and tooling. It aims to simplify the process of configuring, building, and deploying applications by providing predefined templates and automation scripts. The system is designed to reduce manual setup and improve consistency across development environments. It supports modular configurations, allowing users to adapt the build process to different project requirements. The project...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 18
    Make-A-Video - Pytorch (wip)

    Make-A-Video - Pytorch (wip)

    Implementation of Make-A-Video, new SOTA text to video generator

    Implementation of Make-A-Video, new SOTA text to video generator from Meta AI, in Pytorch. They combine pseudo-3d convolutions (axial convolutions) and temporal attention and show much better temporal fusion. The pseudo-3d convolutions isn't a new concept. It has been explored before in other contexts, say for protein contact prediction as "dimensional hybrid residual networks". The gist of the paper comes down to, take a SOTA text-to-image model (here they use DALL-E2, but the same learning points would easily apply to Imagen), make a few minor modifications for attention across time and other ways to skimp on the compute cost, do frame interpolation correctly, get a great video model out. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    SuperPrompt

    SuperPrompt

    Experimental prompt framework exploring reasoning structures in AI

    ...Its main concept revolves around a highly structured prompt format that includes tagged sections for reasoning, analysis, conceptual expansion, and recursive thinking patterns. These sections act as a kind of meta-instruction system intended to influence how an AI model approaches problem solving and conceptual exploration. SuperPrompt also incorporates ideas from mathematics, recursion, abstract algebra, and theoretical reasoning to encourage models to perform multi-step conceptual synthesis rather than simple direct responses.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 20
    Llama Cookbook

    Llama Cookbook

    Solve end to end problems using Llama model family

    The Llama Cookbook is the official Meta LLaMA guide for inference, fine‑tuning, RAG, and multi-step use-cases. It offers recipes, code samples, and integration examples across provider platforms (WhatsApp, SQL, long context workflows), enabling developers to quickly harness LLaMA models
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    TNT

    TNT

    A lightweight library for PyTorch training tools and utilities

    TNT is a lightweight training framework developed by Meta that simplifies the process of building and managing machine learning training loops using PyTorch. The project focuses on providing a flexible yet structured environment for implementing training pipelines without the complexity of large deep learning frameworks. It introduces modular abstractions that allow developers to organize training logic into reusable components such as trainers, evaluators, and callbacks.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 22
    fairseq2

    fairseq2

    FAIR Sequence Modeling Toolkit 2

    fairseq2 is a modern, modular sequence modeling framework developed by Meta AI Research as a complete redesign of the original fairseq library. Built from the ground up for scalability, composability, and research flexibility, fairseq2 supports a broad range of language, speech, and multimodal content generation tasks, including instruction fine-tuning, reinforcement learning from human feedback (RLHF), and large-scale multilingual modeling.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 23
    SkillForge

    SkillForge

    Ultimate meta-skill for generating best-in-class Claude Code skills

    SkillForge is a systematic methodology and tooling framework for creating high-quality AI “skills” specifically optimized for Claude Code integrations, treating skill creation as an engineering discipline rather than an ad-hoc art form. It introduces a multi-phase architecture where every input or request is triaged intelligently, analyzed deeply through structured lenses, specified formally, synthesized with automated generation, and finally subjected to multi-agent review before...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    StyleTTS 2

    StyleTTS 2

    Towards Human-Level Text-to-Speech through Style Diffusion

    ...StyleTTS2 supports both single-speaker and multi-speaker configurations, with the ability to sample or transfer styles from reference audio, making it powerful for expressive TTS and character voices. The repository includes training scripts, configuration files, and pre-trained auxiliary modules such as a text aligner, pitch extractor, and PL-BERT-based linguistic encoder.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 25
    Curated Transformers

    Curated Transformers

    PyTorch library of curated Transformer models and their components

    ...Implementing a feature or bugfix benefits all models. For example, all models support 4/8-bit inference through the bitsandbytes library and each model can use the PyTorch meta device to avoid unnecessary allocations and initialization.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • Next
MongoDB Logo MongoDB