Search Results for "yolov4.weights" - Page 4

Showing 129 open source projects for "yolov4.weights"

View related business solutions
  • Fully Managed MySQL, PostgreSQL, and SQL Server Icon
    Fully Managed MySQL, PostgreSQL, and SQL Server

    Automatic backups, patching, replication, and failover. Focus on your app, not your database.

    Cloud SQL handles your database ops end to end, so you can focus on your app.
    Try Free
  • Go from Code to Production URL in Seconds Icon
    Go from Code to Production URL in Seconds

    Cloud Run deploys apps in any language instantly. Scales to zero. Pay only when code runs.

    Skip the Kubernetes configs. Cloud Run handles HTTPS, scaling, and infrastructure automatically. Two million requests free per month.
    Try it free
  • 1
    InternLM

    InternLM

    Official release of InternLM series

    ...The repository highlights multiple model sizes intended to serve different needs, from efficient research and prototyping to more capable deployments for complex scenarios. Beyond model weights, the project emphasizes an ecosystem view, pointing developers to compatible tools and projects across training and inference so teams can build end-to-end workflows. InternLM’s direction includes strong general-purpose capabilities and ongoing iterations that target improved reasoning, coding, and tool-use behaviors. The broader InternLM ecosystem also includes training tooling and guidance aimed at making fine-tuning and adaptation more accessible across hardware setups, including smaller single-GPU environments and larger multi-node configurations.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    CoTracker

    CoTracker

    CoTracker is a model for tracking any point (pixel) on a video

    ...Its transformer-style architecture aggregates information both along time and across points, allowing it to recover tracks even after brief disappearances. The repository ships with inference scripts, pretrained weights, and simple interfaces to seed points, run tracking, and export trajectories for downstream tasks. Typical uses include correspondence building, motion analysis, dynamic SLAM priors, video editing masks, and evaluation of geometric consistency in real scenes.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Chinese Llama 2 7B

    Chinese Llama 2 7B

    The first Chinese LLaMA2 model in the open source community

    ...The project provides a version of LLaMA-2 that has been further trained on Chinese data so it can better understand and generate text in Chinese while maintaining compatibility with the original model ecosystem. In addition to the model weights, the repository also includes supervised fine-tuning datasets and training resources that help developers build chat-optimized versions of the model. The project follows the input format used by the LLaMA-2 chat architecture, ensuring compatibility with existing optimization techniques and tools built for the LLaMA-2 ecosystem. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Detectron

    Detectron

    FAIR's research platform for object detection research

    Detectron is an object detection and instance segmentation research framework that popularized many modern detection models in a single, reproducible codebase. Built on Caffe2 with custom CUDA/C++ operators, it provided reference implementations for models like Faster R-CNN, Mask R-CNN, RetinaNet, and Feature Pyramid Networks. The framework emphasized a clean configuration system, strong baselines, and a “model zoo” so researchers could compare results under consistent settings. It includes...
    Downloads: 0 This Week
    Last Update:
    See Project
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 5
    LLaMA

    LLaMA

    Inference code for Llama models

    “Llama” is the repository from Meta (formerly Facebook/Meta Research) containing the inference code for LLaMA (Large Language Model Meta AI) models. It provides utilities to load pre-trained LLaMA model weights, run inference (text generation, chat, completions), and work with tokenizers. Tokenizer utilities, download scripts, shell helpers to fetch model weights with correct licensing/permissions. Includes example scripts for chat completions and text completions to show how to call the models in code. This repo is a core piece of the Llama model infrastructure, used by researchers and developers to run LLaMA models locally or in their infrastructure. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    ControlNet

    ControlNet

    Let us control diffusion models

    ControlNet is a neural network architecture designed to add conditional control to text-to-image diffusion models. Rather than training from scratch, ControlNet “locks” the weights of a pre-trained diffusion model and introduces a parallel trainable branch that learns additional conditions—like edges, depth maps, segmentation, human pose, scribbles, or other guidance signals. This allows the system to control where and how the model should focus during generation, enabling users to steer layout, structure, and content more precisely than prompt text alone. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 7
    YoloV3 Implemented in TensorFlow 2.0

    YoloV3 Implemented in TensorFlow 2.0

    YoloV3 Implemented in Tensorflow 2.0

    YoloV3 Implemented in TensorFlow 2.0 is built using TensorFlow 2.0. The project provides a modern deep learning implementation of the popular YOLOv3 algorithm, which is widely used for real-time object detection in images and video streams. YOLOv3 works by dividing an image into grid regions and predicting bounding boxes and class probabilities simultaneously, allowing objects to be detected quickly and efficiently. The repository includes training scripts, inference tools, and configuration...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Stable Diffusion in Docker

    Stable Diffusion in Docker

    Run the Stable Diffusion releases in a Docker container

    Run the Stable Diffusion releases in a Docker container with txt2img, img2img, depth2img, pix2pix, upscale4x, and inpaint. Run the Stable Diffusion releases on Huggingface in a GPU-accelerated Docker container. By default, the pipeline uses the full model and weights which requires a CUDA capable GPU with 8GB+ of VRAM. It should take a few seconds to create one image. On less powerful GPUs you may need to modify some of the options; see the Examples section for more details. If you lack a suitable GPU you can set the options --device cpu and --onnx instead. Since it uses the model, you will need to create a user access token in your Huggingface account. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Detic

    Detic

    Code release for "Detecting Twenty-thousand Classes

    ...The system supports zero- or few-shot extension to novel categories via semantic embeddings and class name supervision, making “open-world” detection practical. Built on Detectron2, the repo includes configs, pretrained weights, and conversion tools to mix fully and weakly supervised sources. Detic is especially useful for applications where label space is vast and long-tailed, but dense bounding-box annotation is infeasible.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • 10
    DiT (Diffusion Transformers)

    DiT (Diffusion Transformers)

    Official PyTorch Implementation of "Scalable Diffusion Models"

    DiT (Diffusion Transformer) is a powerful architecture that applies transformer-based modeling directly to diffusion generative processes for high-quality image synthesis. Unlike CNN-based diffusion models, DiT represents the diffusion process in the latent space and processes image tokens through transformer blocks with learned positional encodings, offering scalability and superior sample quality. The model architecture parallels large language models but for image tokens—each block...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Transformers-Interpret

    Transformers-Interpret

    Model explainability that works seamlessly with Hugging Face

    Transformers-Interpret is an interpretability tool for Transformer-based NLP models, providing insights into attention mechanisms and feature importance.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 12
    Hyperformer

    Hyperformer

    Hypergraph Transformer for Skeleton-based Action Recognition

    This is the official implementation of our paper "Hypergraph Transformer for Skeleton-based Action Recognition." Skeleton-based action recognition aims to recognize human actions given human joint coordinates with skeletal interconnections. By defining a graph with joints as vertices and their natural connections as edges, previous works successfully adopted Graph Convolutional networks (GCNs) to model joint co-occurrences and achieved superior performance. More recently, a limitation of...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    MMGeneration

    MMGeneration

    MMGeneration is a powerful toolkit for generative models

    ...A new design for complex loss modules is proposed for customizing the links between modules, which can achieve flexible combinations among different modules. Conditional GANs have been supported in our toolkit. More methods and pre-trained weights will come soon.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    ConvNeXt V2

    ConvNeXt V2

    Code release for ConvNeXt V2 model

    ...The result is a convnet that competes strongly with transformer architectures on recognition benchmarks while being efficient and hardware-friendly. The repository provides official PyTorch implementations for multiple model sizes (Atto, Femto, Pico, up through Huge), conversion from JAX weights, code for pretraining/fine-tuning, and pretrained checkpoints. It supports both self-supervised pretraining and supervised fine-tuning.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Neural Network Visualization

    Neural Network Visualization

    Project for processing neural networks and rendering to gain insights

    ...It provides an interactive, graphical representation of how data flows through neural network layers, offering a unique educational experience for those new to deep learning or looking to explain it visually. By animating input, weights, activations, and outputs, the tool demystifies neural network operations and helps users intuitively grasp complex concepts. Its lightweight codebase is great for customization and teaching purposes.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Video Pre-Training

    Video Pre-Training

    Learning to Act by Watching Unlabeled Online Videos

    ...The repository contains demonstration models of different widths, fine-tuned variants (e.g. for building houses or early-game tasks), and inference scripts that instantiate agents from pretrained weights. Key modules include the behavioral cloning logic, the agent wrapper, and data loading pipelines (with an accessible skeleton for loading Minecraft demonstration data). The repo also includes a run_agent.py script for testing an agent interactively, and an agent.py module encapsulating the control logic.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    min(DALL·E)

    min(DALL·E)

    min(DALL·E) is a fast, minimal port of DALL·E Mini to PyTorch

    This is a fast, minimal port of Boris Dayma's DALL·E Mini (with mega weights). It has been stripped down for inference and converted to PyTorch. The only third-party dependencies are numpy, requests, pillow and torch. The required models will be downloaded to models_root if they are not already there. Set the dtype to torch.float16 to save GPU memory. If you have an Ampere architecture GPU you can use torch.bfloat16.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    Alphafold2

    Alphafold2

    Unofficial Pytorch implementation / replication of Alphafold2

    ...Once this is replicated, I intend to fold all available amino acid sequences out there in-silico and release it as an academic torrent, to further science. Deepmind has open sourced the official code in Jax, along with the weights! This repository will now be geared towards a straight pytorch translation with some improvements on positional encoding. lhatsk has reported training a modified trunk of this repository, using the same setup as trRosetta, with competitive results. The underlying assumption is that the trunk works on the residue level, and then constitutes to atomic level for the structure module, whether it be SE3 Transformers, E(n)-Transformer, or EGNN doing the refinement.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 19
    YOLOX

    YOLOX

    YOLOX is a high-performance anchor-free YOLO, exceeding yolov3~v5

    ...One more thing worth noting is that you should also implement pull_item and load_anno method for the Mosiac and MixUp augmentations. Except special cases, we always recommend using our COCO pre-trained weights for initializing the model. As YOLOX is an anchor-free detector with only several hyper-parameters, most of the time good results can be obtained with no changes to the models or training settings.
    Downloads: 14 This Week
    Last Update:
    See Project
  • 20
    bert4keras

    bert4keras

    Keras implement of transformers for humans

    ...This is the keras version of the transformer model library re-implemented by the author and is committed to combining transformer and keras with as clean code as possible. The original intention of this project is for the convenience of modification and customization, so it may be updated frequently. Load the pre-trained weights of bert/roberta/albert for fine-tune. Implement the attention mask required by the language model and seq2seq. Pre-training code from zero (supports TPU, multi-GPU, please see pertaining). Compatible with keras, tf.keras.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    Deep Feature Rotation Multimodal Image

    Deep Feature Rotation Multimodal Image

    Implementation of Deep Feature Rotation for Multimodal Image

    Official implementation of paper Deep Feature Rotation for Multimodal Image Style Transfer [NICS'21] We propose a simple method for representing style features in many ways called Deep Feature Rotation (DFR), while still achieving effective stylization compared to more complex methods in style transfer. Our approach is a representative of the many ways of augmentation for intermediate feature embedding without consuming too much computational expense. Prepare your content image and style...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    CodeSearchNet

    CodeSearchNet

    Datasets, tools, and benchmarks for representation learning of code

    CodeSearchNet is a large-scale dataset and research benchmark designed to advance the development of systems that retrieve source code using natural language queries. The project was created through collaboration between GitHub and Microsoft Research and aims to support research on semantic code search and program understanding. The dataset contains millions of pairs of source code functions and corresponding documentation comments extracted from open-source repositories. These pairs allow...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 23
    TRACER

    TRACER

    Extreme Attention Guided Salient Object Tracing Network

    Extreme Attention Guided Salient Object Tracing Network (AAAI 2022) implementation in PyTorch. Now, fast inference mode offers a salient object result with the mask. You can get the more clear salient object by tuning the threshold. We will release initializing TRACER with a version of pre-trained TE-x.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    YOLOv4-large

    YOLOv4-large

    Scaled-YOLOv4: Scaling Cross Stage Partial Network

    YOLOv4-large is an open-source implementation of the Scaled-YOLOv4 object detection architecture, designed to improve both the accuracy and scalability of real-time computer vision models. The project provides a PyTorch implementation of the Scaled-YOLOv4 framework, which extends the original YOLOv4 architecture using Cross Stage Partial (CSP) networks and new scaling techniques.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    CLIP Guided Diffusion

    CLIP Guided Diffusion

    A CLI tool/python module for generating images from text

    A CLI tool/python module for generating images from text using guided diffusion and CLIP from OpenAI. Text to image generation (multiple prompts with weights). Non-square Generations (experimental) Generate portrait or landscape images by specifying a number to offset the width and/or height. Uses fewer timesteps over the same diffusion schedule. Sacrifices accuracy/alignment for quicker runtime. options: - 25, 50, 150, 250, 500, 1000, ddim25,ddim50,ddim150, ddim250,ddim500,ddim1000 (default: 1000) Prepending a number with ddim will use the ddim scheduler. e.g. ddim25 will use the 25 timstep ddim scheduler. ...
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB