Showing 6 open source projects for "ml"

View related business solutions
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build, govern, and optimize agents and models with Gemini Enterprise Agent Platform.
    Start Free
  • Go from Code to Production URL in Seconds Icon
    Go from Code to Production URL in Seconds

    Cloud Run deploys apps in any language instantly. Scales to zero. Pay only when code runs.

    Skip the Kubernetes configs. Cloud Run handles HTTPS, scaling, and infrastructure automatically. Two million requests free per month.
    Try it free
  • 1
    Kubeflow Training Operator

    Kubeflow Training Operator

    Distributed ML Training and Fine-Tuning on Kubernetes

    Kubeflow Training Operator is a Kubernetes-native project for fine-tuning and scalable distributed training of machine learning (ML) models created with various ML frameworks such as PyTorch, TensorFlow, XGBoost, MPI, Paddle, and others.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 2
    Weaviate

    Weaviate

    Weaviate is a cloud-native, modular, real-time vector search engine

    Weaviate in a nutshell: Weaviate is a vector search engine and vector database. Weaviate uses machine learning to vectorize and store data, and to find answers to natural language queries. With Weaviate you can also bring your custom ML models to production scale. Weaviate in detail: Weaviate is a low-latency vector search engine with out-of-the-box support for different media types (text, images, etc.). It offers Semantic Search, Question-Answer-Extraction, Classification, Customizable Models (PyTorch/TensorFlow/Keras), and more. Built from scratch in Go, Weaviate stores both objects and vectors, allowing for combining vector search with structured filtering with the fault-tolerance of a cloud-native database, all accessible through GraphQL, REST, and various language clients.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 3
    Katib

    Katib

    Automated Machine Learning on Kubernetes

    Katib is a Kubernetes-native project for automated machine learning (AutoML). Katib supports Hyperparameter Tuning, Early Stopping and Neural Architecture Search. Katib is a project that is agnostic to machine learning (ML) frameworks. It can tune hyperparameters of applications written in any language of the users’ choice and natively supports many ML frameworks, such as TensorFlow, Apache MXNet, PyTorch, XGBoost, and others. Katib can perform training jobs using any Kubernetes Custom Resources with out-of-the-box support for Kubeflow Training Operator, Argo Workflows, Tekton Pipelines, and many more.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Featureform

    Featureform

    Turn your existing data infrastructure into a feature store

    Featureform allows data scientists to define, manage, and serve machine learning features across your organization. The days of untitled_128.ipynb are over. Transformations, features, and training sets can be pushed from notebooks to a centralized feature repository with metadata like name, variant, lineage, and owner. Featureform's Virtual Feature Store architecture orchestrates your data infrastructure to build and maintain your training sets and production features. It offers a framework...
    Downloads: 5 This Week
    Last Update:
    See Project
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 5
    Gorgonia

    Gorgonia

    Gorgonia is a library that helps facilitate machine learning in Go

    ...The primary goal for Gorgonia is to be a highly performant machine learning/graph computation-based library that can scale across multiple machines. It should bring the appeal of Go (simple compilation and deployment process) to the ML world. It's a long way from there currently, however, the baby steps are already there. The main reason to use Gorgonia is developer comfort. If you're using a Go stack extensively, now you have access to the ability to create production-ready machine learning systems in an environment that you are already familiar and comfortable with.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 6
    cortex

    cortex

    Production infrastructure for machine learning at scale

    ...Cortex handles many operational challenges associated with deploying AI systems, such as managing dependencies, orchestrating data pipelines, and scaling services under load. Developers can define machine learning pipelines as code using declarative configuration files, which simplifies the process of managing complex ML workflows. The platform supports integration with cloud environments and container orchestration systems so that applications can scale dynamically based on demand. It is designed to help teams focus on building machine learning logic rather than managing infrastructure details.
    Downloads: 2 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB