62 Integrations with Hugging Face

View a list of Hugging Face integrations and software that integrates with Hugging Face below. Compare the best Hugging Face integrations as well as features, ratings, user reviews, and pricing of software that integrates with Hugging Face. Here are the current Hugging Face integrations in 2024:

  • 1
    Anycode AI

    Anycode AI

    Anycode AI

    The only auto-pilot agent that works with your unique software development workflow. Anycode AI converts your whole Legacy codebase to modern tech stacks up to 8X faster. Boost your coding speed tenfold with Anycode AI. Utilize AI for rapid, compliant coding and testing. Modernize swiftly with Anycode AI. Effortlessly handle legacy code and embrace updates for efficient applications. Upgrade seamlessly from outdated systems. Our platform refines old logic for a smooth transition to advanced tech.
  • 2
    LM Studio

    LM Studio

    LM Studio

    Use models through the in-app Chat UI or an OpenAI-compatible local server. Minimum requirements: M1/M2/M3 Mac, or a Windows PC with a processor that supports AVX2. Linux is available in beta. One of the main reasons for using a local LLM is privacy, and LM Studio is designed for that. Your data remains private and local to your machine. You can use LLMs you load within LM Studio via an API server running on localhost.
  • 3
    3LC

    3LC

    3LC

    Light up the black box and pip install 3LC to gain the clarity you need to make meaningful changes to your models in moments. Remove the guesswork from your model training and iterate fast. Collect per-sample metrics and visualize them in your browser. Analyze your training and eliminate issues in your dataset. Model-guided, interactive data debugging and enhancements. Find important or inefficient samples. Understand what samples work and where your model struggles. Improve your model in different ways by weighting your data. Make sparse, non-destructive edits to individual samples or in a batch. Maintain a lineage of all changes and restore any previous revisions. Dive deeper than standard experiment trackers with per-sample per epoch metrics and data tracking. Aggregate metrics by sample features, rather than just epoch, to spot hidden trends. Tie each training run to a specific dataset revision for full reproducibility.
  • 4
    EvalsOne

    EvalsOne

    EvalsOne

    An intuitive yet comprehensive evaluation platform to iteratively optimize your AI-driven products. Streamline LLMOps workflow, build confidence, and gain a competitive edge. EvalsOne is your all-in-one toolbox for optimizing your application evaluation process. Imagine a Swiss Army knife for AI, equipped to tackle any evaluation scenario you throw its way. Suitable for crafting LLM prompts, fine-tuning RAG processes, and evaluating AI agents. Choose from rule-based or LLM-based approaches to automate the evaluation process. Integrate human evaluation seamlessly, leveraging the power of expert judgment. Applicable to all LLMOps stages from development to production environments. EvalsOne provides an intuitive process and interface, that empowers teams across the AI lifecycle, from developers to researchers and domain experts. Easily create evaluation runs and organize them in levels. Quickly iterate and perform in-depth analysis through forked runs.
  • 5
    Gemma 2

    Gemma 2

    Google

    A family of state-of-the-art, light-open models created from the same research and technology that were used to create Gemini models. These models incorporate comprehensive security measures and help ensure responsible and reliable AI solutions through selected data sets and rigorous adjustments. Gemma models achieve exceptional comparative results in their 2B, 7B, 9B, and 27B sizes, even outperforming some larger open models. With Keras 3.0, enjoy seamless compatibility with JAX, TensorFlow, and PyTorch, allowing you to effortlessly choose and change frameworks based on task. Redesigned to deliver outstanding performance and unmatched efficiency, Gemma 2 is optimized for incredibly fast inference on various hardware. The Gemma family of models offers different models that are optimized for specific use cases and adapt to your needs. Gemma models are large text-to-text lightweight language models with a decoder, trained in a huge set of text data, code, and mathematical content.
  • 6
    Codestral Mamba
    As a tribute to Cleopatra, whose glorious destiny ended in tragic snake circumstances, we are proud to release Codestral Mamba, a Mamba2 language model specialized in code generation, available under an Apache 2.0 license. Codestral Mamba is another step in our effort to study and provide new architectures. It is available for free use, modification, and distribution, and we hope it will open new perspectives in architecture research. Mamba models offer the advantage of linear time inference and the theoretical ability to model sequences of infinite length. It allows users to engage with the model extensively with quick responses, irrespective of the input length. This efficiency is especially relevant for code productivity use cases, this is why we trained this model with advanced code and reasoning capabilities, enabling it to perform on par with SOTA transformer-based models.
  • 7
    Mathstral

    Mathstral

    Mistral AI

    As a tribute to Archimedes, whose 2311th anniversary we’re celebrating this year, we are proud to release our first Mathstral model, a specific 7B model designed for math reasoning and scientific discovery. The model has a 32k context window published under the Apache 2.0 license. We’re contributing Mathstral to the science community to bolster efforts in advanced mathematical problems requiring complex, multi-step logical reasoning. The Mathstral release is part of our broader effort to support academic projects, it was produced in the context of our collaboration with Project Numina. Akin to Isaac Newton in his time, Mathstral stands on the shoulders of Mistral 7B and specializes in STEM subjects. It achieves state-of-the-art reasoning capacities in its size category across various industry-standard benchmarks. In particular, it achieves 56.6% on MATH and 63.47% on MMLU, with the following MMLU performance difference by subject between Mathstral 7B and Mistral 7B.
  • 8
    Jamba

    Jamba

    AI21 Labs

    Jamba is the most powerful & efficient long context model, open for builders and built for the enterprise. Jamba's latency outperforms all leading models of comparable sizes. Jamba's 256k context window is the longest openly available. Jamba's Mamba-Transformer MoE architecture is designed for cost & efficiency gains. Jamba offers key features of OOTB including function calls, JSON mode output, document objects, and citation mode. Jamba 1.5 models maintain high performance across the full length of their context window. Jamba 1.5 models achieve top scores across common quality benchmarks. Secure deployment that suits your enterprise. Seamlessly start using Jamba on our production-grade SaaS platform. The Jamba model family is available for deployment across our strategic partners. We offer VPC & on-prem deployments for enterprises that require custom solutions. For enterprises that have unique, bespoke requirements, we offer hands-on management, continuous pre-training, etc.
  • 9
    crewAI+

    crewAI+

    crewAI

    Turn any crew into an API within seconds. Connect to your apps using hooks, REST, gRPC, and more. Get access to templates, custom tools, and early UI. Get business support, SLA, and private VPC. We provide the power with simplicity. We provide a platform and hope you create wonders. Cutting-edge framework for orchestrating autonomous AI agents. By fostering collaborative intelligence, CrewAI empowers agents to work together seamlessly, tackling complex tasks. Use our developer framework and tools to build your own multi-agent automation locally. Become an early beta customer of crewAI+, get access to early features, get a discount, and help shape it.
  • 10
    Acuvity

    Acuvity

    Acuvity

    Acuvity is the most comprehensive AI security and governance platform for your employees and applications. DevSecOps implements AI security without code changes and devs can focus on AI Innovation. Pluggable AI security results in completeness of coverage, without old libraries or insufficient coverage. Optimize costs by efficiently using GPUs only for LLM models. Full visibility into all GenAI models, apps, plugins, and services that your teams are using and exploring. Granular observability into all GenAI interactions with comprehensive logging and an audit trail of inputs and outputs. AI usage in enterprises requires a specialized security framework that is able to address new AI risk vectors and comply with emerging AI regulations. Employees can use AI confidently, without risking exposing confidential data. Legal would like to ensure there are no copyright, or regulatory issues while using AI-generated content.
  • 11
    Outspeed

    Outspeed

    Outspeed

    Outspeed provides networking and inference infrastructure to build fast, real-time voice and video AI apps. AI-powered speech recognition, natural language processing, and text-to-speech for intelligent voice assistants, automated transcription, and voice-controlled systems. Create interactive digital characters for virtual hosts, AI tutors, or customer service. Enable real-time animation and natural conversations for engaging digital interactions. Real-time visual AI for quality control, surveillance, touchless interactions, and medical imaging analysis. Process and analyze video streams and images with high speed and accuracy. AI-driven content generation for creating vast, detailed digital worlds efficiently. Ideal for game environments, architectural visualizations, and virtual reality experiences. Create custom multimodal AI solutions with Adapt's flexible SDK and infrastructure. Combine AI models, data sources, and interaction modes for innovative applications.
  • 12
    Unremot

    Unremot

    Unremot

    Unremot is a go-to place for anyone aspiring to build an AI product - with 120+ pre-built APIs, you can build and launch AI products 2X faster, at 1/3rd cost. Even, some of the most complicated AI product APIs take less than a few minutes to deploy and launch, with minimal code or even no-code. Choose an AI API that you want to integrate to your product from 120+ APIs we have on Unremot. Provide your API private key to authenticate Unremot to access the API. Use unremot unique URL to connect the product API - the whole process takes only minutes, instead of days and weeks.