Showing 698 open source projects for "tuning"

View related business solutions
  • Go From AI Idea to AI App Fast Icon
    Go From AI Idea to AI App Fast

    One platform to build, fine-tune, and deploy ML models. No MLOps team required.

    Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
    Try Free
  • Auth0 B2B Essentials: SSO, MFA, and RBAC Built In Icon
    Auth0 B2B Essentials: SSO, MFA, and RBAC Built In

    Unlimited organizations, 3 enterprise SSO connections, role-based access control, and pro MFA included. Dev and prod tenants out of the box.

    Auth0's B2B Essentials plan gives you everything you need to ship secure multi-tenant apps. Unlimited orgs, enterprise SSO, RBAC, audit log streaming, and higher auth and API limits included. Add on M2M tokens, enterprise MFA, or additional SSO connections as you scale.
    Sign Up Free
  • 1
    LLaMA Efficient Tuning

    LLaMA Efficient Tuning

    Easy-to-use LLM fine-tuning framework (LLaMA-2, BLOOM, Falcon

    Easy-to-use LLM fine-tuning framework (LLaMA-2, BLOOM, Falcon, Baichuan, Qwen, ChatGLM2)
    Downloads: 5 This Week
    Last Update:
    See Project
  • 2
    Universal x86 Tuning Utility

    Universal x86 Tuning Utility

    Unlock the full potential of your Intel/AMD based device

    Unlock the full potential of your Intel/AMD-based device.
    Downloads: 18 This Week
    Last Update:
    See Project
  • 3
    PEFT

    PEFT

    State-of-the-art Parameter-Efficient Fine-Tuning

    Parameter-Efficient Fine-Tuning (PEFT) methods enable efficient adaptation of pre-trained language models (PLMs) to various downstream applications without fine-tuning all the model's parameters. Fine-tuning large-scale PLMs is often prohibitively costly. In this regard, PEFT methods only fine-tune a small number of (extra) model parameters, thereby greatly decreasing the computational and storage costs.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    LLaMA-Factory

    LLaMA-Factory

    Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)

    LLaMA-Factory is a fine-tuning and training framework for Meta's LLaMA language models. It enables researchers and developers to train and customize LLaMA models efficiently using advanced optimization techniques.
    Downloads: 12 This Week
    Last Update:
    See Project
  • Forever Free Full-Stack Observability | Grafana Cloud Icon
    Forever Free Full-Stack Observability | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 5
    SimpleTuner

    SimpleTuner

    A general fine-tuning kit geared toward image/video/audio diffusion

    SimpleTuner is an open-source toolkit designed to simplify the fine-tuning of modern diffusion models for generating images, video, and audio. The project focuses on providing a clear and understandable training environment for researchers, developers, and artists who want to customize generative AI models without navigating complex machine learning pipelines. It supports fine-tuning workflows for models such as Stable Diffusion variants and other diffusion architectures, enabling users to adapt pretrained models to specialized datasets or creative tasks. ...
    Downloads: 8 This Week
    Last Update:
    See Project
  • 6
    SetFit

    SetFit

    Efficient few-shot learning with Sentence Transformers

    SetFit is an efficient and prompt-free framework for few-shot fine-tuning of Sentence Transformers. It achieves high accuracy with little labeled data - for instance, with only 8 labeled examples per class on the Customer Reviews sentiment dataset, SetFit is competitive with fine-tuning RoBERTa Large on the full training set of 3k examples.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 7
    Gemma in PyTorch

    Gemma in PyTorch

    The official PyTorch implementation of Google's Gemma models

    gemma_pytorch provides the official PyTorch reference for running and fine-tuning Google’s Gemma family of open models. It includes model definitions, configuration files, and loading utilities for multiple parameter scales, enabling quick evaluation and downstream adaptation. The repository demonstrates text generation pipelines, tokenizer setup, quantization paths, and adapters for low-rank or parameter-efficient fine-tuning.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Llama Stack

    Llama Stack

    Composable building blocks to build Llama Apps

    Llama-Stack is an open-source framework designed to facilitate the deployment and fine-tuning of large language models (LLMs) for various natural language processing tasks.
    Downloads: 10 This Week
    Last Update:
    See Project
  • 9
    qvac-fabric-llm.cpp

    qvac-fabric-llm.cpp

    QVAC Fabric: cross-platform LLM inference and fine-tuning

    ...It introduces native LoRA fine-tuning capabilities that can be executed directly on consumer hardware, allowing developers to train and adapt models locally without relying on cloud infrastructure. A key innovation is its support for BitNet ternary quantized models, enabling highly efficient inference and training even on resource-constrained systems.
    Downloads: 2 This Week
    Last Update:
    See Project
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 10
    PaLM + RLHF - Pytorch

    PaLM + RLHF - Pytorch

    Implementation of RLHF (Reinforcement Learning with Human Feedback)

    PaLM-rlhf-pytorch is a PyTorch implementation of Pathways Language Model (PaLM) with Reinforcement Learning from Human Feedback (RLHF). It is designed for fine-tuning large-scale language models with human preference alignment, similar to OpenAI’s approach for training models like ChatGPT.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 11
    InstantCharacter

    InstantCharacter

    Personalize Any Characters with a Scalable Diffusion Transformer

    InstantCharacter is a tuning-free diffusion transformer framework created by Tencent Hunyuan / InstantX team, which enables generating images of a specific character (subject) from a single reference image, preserving identity and character features. Uses adapters, so full fine-tuning of the base model is not required. Demo scripts and pipeline API (via infer_demo.py, pipeline.py) included.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Pyreft

    Pyreft

    ReFT: Representation Finetuning for Language Models

    PyreFT is a tool by Stanford NLP for fine-tuning transformer models with an emphasis on efficient, resource-conserving training and customizability for NLP tasks.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 13
    Huatuo-Llama-Med-Chinese

    Huatuo-Llama-Med-Chinese

    Instruction-tuning LLM with Chinese Medical Knowledge

    Huatuo-Llama-Med-Chinese is an open-source project that develops medical-domain large language models by instruction-tuning existing models using Chinese medical knowledge. The project builds specialized models by fine-tuning architectures such as LLaMA, Alpaca-Chinese, and Bloom with curated medical datasets. These datasets are constructed from medical knowledge graphs, academic literature, and question-answer pairs designed to teach models how to respond accurately to healthcare-related queries. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Gemma

    Gemma

    Gemma open-weight LLM library, from Google DeepMind

    ...The framework supports both text and multi-modal input, allowing natural language conversations that incorporate visual content such as images. It includes APIs for conversational sampling, parameter management, and integration with fine-tuning methods like LoRA. The Gemma library can operate efficiently on CPUs, GPUs, or TPUs, with recommended configurations depending on model size. Through included tutorials and Colab notebooks, users can explore examples covering sampling, multi-modal interactions, and fine-tuning workflows. By providing accessible open-weight models, Gemma enables researchers and developers to experiment with state-of-the-art LLM architectures.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 15
    VisualGLM-6B

    VisualGLM-6B

    Chinese and English multimodal conversational language model

    ...Trained on a large bilingual dataset — including 30 million high-quality Chinese image-text pairs from CogView and 300 million English pairs — VisualGLM-6B is designed for image understanding, description, and question answering. Fine-tuning on long visual QA datasets further aligns the model’s responses with human preferences. The repository provides inference APIs, command-line demos, web demos, and efficient fine-tuning options like LoRA, QLoRA, and P-tuning. It also supports quantization down to INT4, enabling local deployment on consumer GPUs with as little as 6.3 GB VRAM.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 16
    Hands-On Large Language Models

    Hands-On Large Language Models

    Official code repo for the O'Reilly Book

    ...The repository is structured into chapters that align with the educational progression of the book — covering everything from foundational topics like tokens, embeddings, and transformer architecture to advanced techniques such as prompt engineering, semantic search, retrieval-augmented generation (RAG), multimodal LLMs, and fine-tuning. Each chapter contains executable Jupyter notebooks that are designed to be run in environments like Google Colab, making it easy for learners to experiment interactively with models, visualize attention patterns, implement classification and generation tasks.
    Downloads: 176 This Week
    Last Update:
    See Project
  • 17
    Easy DataSet

    Easy DataSet

    A powerful tool for creating datasets for LLM fine-tuning

    Easy DataSet is a comprehensive open-source tool designed to make creating high-quality datasets for large language model fine-tuning, retrieval-augmented generation (RAG), and evaluation as easy and automated as possible by providing intuitive interfaces and powerful parsing, segmentation, and labeling tools. It supports ingesting domain-specific documents in a wide range of formats — including PDF, Markdown, DOCX, EPUB, and plain text — and can intelligently segment, clean, and structure content into rich datasets tailored for downstream LLM training needs. ...
    Downloads: 12 This Week
    Last Update:
    See Project
  • 18
    Kubeflow Training Operator

    Kubeflow Training Operator

    Distributed ML Training and Fine-Tuning on Kubernetes

    Kubeflow Training Operator is a Kubernetes-native project for fine-tuning and scalable distributed training of machine learning (ML) models created with various ML frameworks such as PyTorch, TensorFlow, XGBoost, MPI, Paddle, and others.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 19
    FLAML

    FLAML

    A fast library for AutoML and tuning

    ...Users can find their desired customizability from a smooth range: minimal customization (computational resource budget), medium customization (e.g., scikit-style learner, search space, and metric), or full customization (arbitrary training and evaluation code). It supports fast automatic tuning, capable of handling complex constraints/guidance/early stopping. FLAML is powered by a new, cost-effective hyperparameter optimization and learner selection method invented by Microsoft Research.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 20
    NVIDIA Profile Inspector

    NVIDIA Profile Inspector

    Modify game profiles inside the internal driver database

    ...It exposes advanced and undocumented configuration options that can influence rendering behavior, performance optimization, and compatibility for specific games. Users can create, edit, and assign profiles for individual applications, enabling fine-grained tuning of GPU behavior beyond standard settings. The tool is particularly popular among enthusiasts who want to optimize performance, troubleshoot graphical issues, or enable experimental features such as custom DLSS configurations. It provides a detailed interface for browsing and editing driver parameters, making it possible to adjust settings that are otherwise inaccessible.
    Downloads: 52 This Week
    Last Update:
    See Project
  • 21
    Transformers.jl

    Transformers.jl

    Julia Implementation of Transformer models

    ...Inspired by architectures like BERT, GPT, and T5, the library offers a modular and flexible interface for building, training, and using transformer-based deep learning models. It supports training from scratch and fine-tuning pretrained models, and integrates with Flux.jl for automatic differentiation and optimization.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 22
    BayesianOptimization

    BayesianOptimization

    A Python implementation of global optimization with gaussian processes

    BayesianOptimization is a Python library that helps find the maximum (or minimum) of expensive or unknown objective functions using Bayesian optimization. This technique is especially useful for hyperparameter tuning in machine learning, where evaluating the objective function is costly. The library provides an easy-to-use API for defining bounds and optimizing over parameter spaces using probabilistic models like Gaussian Processes.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 23
    Axolotl

    Axolotl

    Go ahead and axolotl questions

    Axolotl is a powerful and flexible framework for fine-tuning large language models on custom datasets. Built for researchers and developers, Axolotl simplifies the process of adapting LLMs for specific tasks, including chat, code generation, and instruction following. It supports a wide variety of model architectures and offers out-of-the-box optimization strategies for efficient training.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 24
    Lazy Predict

    Lazy Predict

    Lazy Predict help build a lot of basic models without much code

    Lazy Predict helps build a lot of basic models without much code and helps understand which models work better without any parameter tuning.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 25
    Xray Core

    Xray Core

    Xray, Penetrates Everything. Also the best v2ray-core

    Xray-core is an enhanced superset of v2ray-core, providing a high-performance privacy proxy framework supporting XTLS. It offers full compatibility with v2ray configs, supports multiple protocols, advanced routing, and is distributed as a single executable.
    Downloads: 459 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • 5
  • Next
MongoDB Logo MongoDB