Search Results for "yolov4.weights" - Page 5

Showing 129 open source projects for "yolov4.weights"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Enterprise-grade ITSM, for every business Icon
    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity.

    Freshservice is an intuitive, AI-powered platform that helps IT, operations, and business teams deliver exceptional service without the usual complexity. Automate repetitive tasks, resolve issues faster, and provide seamless support across the organization. From managing incidents and assets to driving smarter decisions, Freshservice makes it easy to stay efficient and scale with confidence.
    Try it Free
  • 1
    TimeSformer

    TimeSformer

    The official pytorch implementation of our paper

    TimeSformer is a vision transformer architecture for video that extends the standard attention mechanism into spatiotemporal attention. The model alternates attention along spatial and temporal dimensions (or designs variants like divided attention) so that it can capture both appearance and motion cues in video. Because the attention is global across frames, TimeSformer can reason about dependencies across long time spans, not just local neighborhoods. The official implementation in PyTorch...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 2
    PyCls

    PyCls

    Codebase for Image Classification Research, written in PyTorch

    ...Distributed training and mixed precision are first-class, enabling fast experiments on multi-GPU setups with simple, declarative configs. Model definitions are concise and modular, making it easy to prototype new blocks or swap backbones while keeping the rest of the pipeline unchanged. Pretrained weights and evaluation scripts cover common datasets, and the logging/metric stack is designed for quick comparison across runs. Practitioners use pycls both as a baseline factory and as a scaffold for new classification backbones.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Minkowski Engine

    Minkowski Engine

    Auto-diff neural network library for high-dimensional sparse tensors

    ...Compressing a neural network to speed up inference and minimize memory footprint has been studied widely. One of the popular techniques for model compression is pruning the weights in convnets, is also known as sparse convolutional networks. Such parameter-space sparsity used for model compression compresses networks that operate on dense tensors and all intermediate activations of these networks are also dense tensors.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    TransPose

    TransPose

    PyTorch Implementation for "TransPose, Keypoint localization

    TransPose is a human pose estimation model based on a CNN feature extractor, a Transformer Encoder, and a prediction head. Given an image, the attention layers built in Transformer can efficiently capture long-range spatial relationships between keypoints and explain what dependencies the predicted keypoints locations highly rely on.
    Downloads: 2 This Week
    Last Update:
    See Project
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 5
    Keepsake

    Keepsake

    Version control for machine learning

    Keepsake is a Python library that uploads files and metadata (like hyperparameters) to Amazon S3 or Google Cloud Storage. You can get the data back out using the command-line interface or a notebook.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    PyTorch SimCLR

    PyTorch SimCLR

    PyTorch implementation of SimCLR: A Simple Framework

    ...And most important, their features are known to adapt well to new problems. This is particularly interesting when annotated training data is scarce. In situations like this, we take the models’ pre-trained weights, append a new classifier layer on top of it, and retrain the network. This is called transfer learning, and is one of the most used techniques in CV. Aside from a few tricks when performing fine-tuning (if the case), it has been shown (many times) that if training for a new task, models initialized with pre-trained weights tend to learn faster and be more accurate then training from scratch using random initialization.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Phishing Catcher

    Phishing Catcher

    Real-time phishing domain detection via Certificate Transparency logs

    phishing_catcher is a security monitoring tool designed to detect potential phishing domains in near real time by analyzing TLS certificate issuance events. It listens to Certificate Transparency (CT) logs through the CertStream API and evaluates newly issued certificates as they appear. Each certificate often contains one or more domain names, which the tool analyzes to determine whether they resemble suspicious or phishing-related domains. phishing_catcher applies a configurable scoring...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 8
    TFLearn

    TFLearn

    Deep learning library featuring a higher-level API for TensorFlow

    ...Powerful helper functions to train any TensorFlow graph, with support of multiple inputs, outputs, and optimizers. Easy and beautiful graph visualization, with details about weights, gradients, activations, and more. Effortless device placement for using multiple CPU/GPU. The high-level API currently supports the most of the recent deep learning models, such as Convolutions, LSTM, BiRNN, BatchNorm, etc.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Age and Gender Estimation

    Age and Gender Estimation

    Keras implementation of a CNN network for age and gender estimation

    ...In training, the IMDB-WIKI dataset is used. Because the face images in the UTKFace dataset is tightly cropped (there is no margin around the face region), faces should also be cropped in demo.py if weights trained by the UTKFace dataset is used. Please set the margin argument to 0 for tight cropping. You can evaluate a trained model on the APPA-REAL (validation) dataset. We pose the age regression problem as a deep classification problem followed by a softmax expected value refinement and show improvements over direct regression training of CNNs. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • 10
    StellarGraph

    StellarGraph

    Machine Learning on Graphs

    StellarGraph is a Python library for machine learning on graphs and networks. The StellarGraph library offers state-of-the-art algorithms for graph machine learning, making it easy to discover patterns and answer questions about graph-structured data. It can solve many machine learning tasks. Graph-structured data represent entities as nodes (or vertices) and relationships between them as edges (or links), and can include data associated with either as attributes. For example, a graph can...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    ENAS in PyTorch

    ENAS in PyTorch

    PyTorch implementation of "Efficient Neural Architecture Search

    ...The project includes training scripts, model definitions, and search procedures that show the full workflow from architecture sampling to evaluation. Because ENAS relies on shared weights among candidate models, the implementation emphasizes efficiency and experiment reproducibility.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Tensorflow and deep learning

    Tensorflow and deep learning

    A crash course in six episodes for software developers

    ...It is structured as a series of guided lessons that combine theoretical explanations, practical examples, and runnable code, allowing learners to build intuition while actively experimenting with models. The repository covers core neural network concepts such as weights, biases, activation functions, and gradient descent, as well as more advanced techniques like convolutional networks, recurrent networks, and reinforcement learning. It includes multiple hands-on projects, such as handwritten digit recognition, airplane detection in images, and text generation using recurrent neural networks, which demonstrate how different architectures solve real-world problems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    TensorNets

    TensorNets

    High level network definitions with pre-trained weights in TensorFlow

    ...Also, it is easy to deploy and expand a collection of pre-processing and pre-trained weights. Readability. With recent TensorFlow APIs, more factoring and less indenting can be possible. For example, all the inception variants are implemented as about 500 lines of code in TensorNets while 2000+ lines in official TensorFlow models. Reproducibility. You can always reproduce the original results with simple APIs including feature extractions.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14

    PBTK Optimizer

    Application for optimization of parameters in PBTK models

    ...Some limitations exist in the PBTK field, such as difficulty of model development and a lack of application specific software tools to help modelers. Some parameters used in PBTK models, such as tissue weights, are easily measure. Other parameters can be determined through in-vitro experiments or through extrapolation using published equations. When it is impractical to use these methods to estimate a parameter, techniques can be used to optimize parameters so that model results best fit validation data. This tool was designed to optimize a user-specified list of parameters to a user-specified PBTK model. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Bangla TTS

    Bangla TTS

    Bangla text to speech synthesis in python

    ...Installation -------------------------------------- * Install Anaconda * conda create -n new_virtual_env python==3.6.8 * conda activate new_virtual_env * pip install -r requirements.txt * While running for the first time, keep your internet connection on to download the weights of the speech synthesis models (>500 MB) * For fast inference, you must install tensorflow-gpu and have a NVidia GPU. Project link: https://github.com/zabir-nabil/bangla-tts
    Downloads: 1 This Week
    Last Update:
    See Project
  • 16
    textgenrnn

    textgenrnn

    Easily train your own text-generating neural network

    With textgenrnn you can easily train your own text-generating neural network of any size and complexity on any text dataset with a few lines of code. A modern neural network architecture that utilizes new techniques as attention-weighting and skip-embedding to accelerate training and improve model quality. Train on and generate text at either the character-level or word-level. Configure RNN size, the number of RNN layers, and whether to use bidirectional RNNs. Train on any generic input text...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Image Super-Resolution (ISR)

    Image Super-Resolution (ISR)

    Super-scale your images and run experiments with Residual Dense

    ...When training your own model, start with only PSNR loss (50+ epochs, depending on the dataset) and only then introduce GANS and feature loss. This can be controlled by the loss weights argument. The weights used to produce these images are available directly when creating the model object. ISR is compatible with Python 3.6 and is distributed under the Apache 2.0 license.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 18
    gpt2-client

    gpt2-client

    Easy-to-use TensorFlow Wrapper for GPT-2 117M, 345M, 774M, etc.

    GPT-2 is a Natural Language Processing model developed by OpenAI for text generation. It is the successor to the GPT (Generative Pre-trained Transformer) model trained on 40GB of text from the internet. It features a Transformer model that was brought to light by the Attention Is All You Need paper in 2017. The model has 4 versions - 124M, 345M, 774M, and 1558M - that differ in terms of the amount of training data fed to it and the number of parameters they contain. Finally, gpt2-client is a...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    Torchreid

    Torchreid

    Deep learning person re-identification in PyTorch

    ...The folder "configs/" contains some predefined configs which you can use as a starting point. The code will automatically (download and) load the ImageNet pretrained weights. After the training is done, the model will be saved as "log/osnet_x1_0_market1501_softmax_cosinelr/model.pth.tar-250". Under the same folder, you can find the tensorboard file. Different from the same-domain setting, here we replace random_erase with color_jitter. This can improve the generalization performance on the unseen target dataset.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    ELI5

    ELI5

    A library for debugging/inspecting machine learning classifiers

    ...The project focuses on improving model transparency by providing tools that visualize feature importance and prediction reasoning. It supports several popular machine learning frameworks including scikit-learn, XGBoost, LightGBM, CatBoost, and Keras. The library allows users to inspect model weights, analyze decision trees, and compute permutation feature importance for black-box models. It also provides specialized tools such as TextExplainer, which can highlight important words in text classification tasks to explain why a model produced a particular prediction. Additionally, the library integrates explanation algorithms such as LIME to interpret predictions from arbitrary machine learning models.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    FID score for PyTorch

    FID score for PyTorch

    Compute FID scores with PyTorch

    ...It was shown to correlate well with human judgement of visual quality and is most often used to evaluate the quality of samples of Generative Adversarial Networks. FID is calculated by computing the Fréchet distance between two Gaussians fitted to feature representations of the Inception network. The weights and the model are exactly the same as in the official Tensorflow implementation, and were tested to give very similar results (e.g. .08 absolute error and 0.0009 relative error on LSUN, using ProGAN generated images). However, due to differences in the image interpolation implementation and library backends, FID results still differ slightly from the original implementation. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 22
    CakeChat

    CakeChat

    CakeChat: Emotional Generative Dialog System

    ...Decoder can be conditioned on any categorical label, for example, emotion label or persona id. May be initialized using w2v model trained on your corpus. Embedding layer may be either fixed or fine-tuned along with other weights of the network.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    PyTorch pretrained BigGAN

    PyTorch pretrained BigGAN

    PyTorch implementation of BigGAN with pretrained weights

    ...This reimplementation was done from the raw computation graph of the Tensorflow version and behave similarly to the TensorFlow version (variance of the output difference of the order of 1e-5). This implementation currently only contains the generator as the weights of the discriminator were not released (although the structure of the discriminator is very similar to the generator so it could be added pretty easily.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    SSD Keras

    SSD Keras

    A Keras port of single shot MultiBox detector

    This is a Keras port of the SSD model architecture introduced by Wei Liu et al. in the paper SSD: Single Shot MultiBox Detector. Ports of the trained weights of all the original models are provided below. This implementation is accurate, meaning that both the ported weights and models trained from scratch produce the same mAP values as the respective models of the original Caffe implementation. The main goal of this project is to create an SSD implementation that is well documented for those who are interested in a low-level understanding of the model. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 25
    Image classification models for Keras

    Image classification models for Keras

    Keras code and weights files for popular deep learning models

    ...For instance, if you have set image_dim_ordering=tf, then any model loaded from this repository will get built according to the TensorFlow dimension ordering convention, "Width-Height-Depth". Pre-trained weights can be automatically loaded upon instantiation (weights='imagenet' argument in model constructor for all image models, weights='msd' for the music tagging model). Weights are automatically downloaded if necessary, and cached locally in ~/.keras/models/. This repository contains code for the following Keras models, VGG16, VGG19, ResNet50, Inception v3, and CRNN for music tagging.
    Downloads: 2 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB