Showing 29 open source projects for "yolov4.weights"

View related business solutions
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • Enterprise-grade ITSM, for every business Icon
    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity.

    Freshservice is an intuitive, AI-powered platform that helps IT, operations, and business teams deliver exceptional service without the usual complexity. Automate repetitive tasks, resolve issues faster, and provide seamless support across the organization. From managing incidents and assets to driving smarter decisions, Freshservice makes it easy to stay efficient and scale with confidence.
    Try it Free
  • 1
    Weights and Biases

    Weights and Biases

    Tool for visualizing and tracking your machine learning experiments

    ...Capture dataset versions with W&B Artifacts to identify how changing data affects your resulting models. Reproduce any model, with saved code, hyperparameters, launch commands, input data, and resulting model weights. Set wandb.config once at the beginning of your script to save your hyperparameters, input settings (like dataset name or model type), and any other independent variables for your experiments. This is useful for analyzing your experiments and reproducing your work in the future. Setting configs also allows you to visualize the relationships between features of your model architecture or data pipeline and model performance.
    Downloads: 10 This Week
    Last Update:
    See Project
  • 2
    BitNet

    BitNet

    BitNet: Scaling 1-bit Transformers for Large Language Models

    ...The project implements the BitNet architecture described in research on scaling transformer models using extremely low-bit quantization techniques. In this approach, neural network weights are quantized to approximately one bit per parameter, allowing models to operate with far lower memory usage than traditional 16-bit or 32-bit neural networks. The architecture introduces specialized layers such as BitLinear, which replace standard linear projections in transformer networks with quantized operations. By limiting weight precision while maintaining efficient scaling and normalization strategies, the architecture aims to retain competitive performance while significantly reducing hardware requirements.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    EasyOCR

    EasyOCR

    Ready-to-use OCR with 80+ supported languages

    ...Second-generation models: multiple times smaller size, multiple times faster inference, additional characters and comparable accuracy to the first generation models. EasyOCR will choose the latest model by default but you can also specify which model to use. Model weights for the chosen language will be automatically downloaded or you can download them manually from the model hub. The idea is to be able to plug-in any state-of-the-art model into EasyOCR. There are a lot of geniuses trying to make better detection/recognition models, but we are not trying to be geniuses here. We just want to make their works quickly accessible to the public.
    Downloads: 38 This Week
    Last Update:
    See Project
  • 4
    pycm

    pycm

    Multi-class confusion matrix library in Python

    PyCM is a multi-class confusion matrix library written in Python that supports both input data vectors and direct matrix, and a proper tool for post-classification model evaluation that supports most classes and overall statistics parameters. PyCM is the swiss-army knife of confusion matrices, targeted mainly at data scientists that need a broad array of metrics for predictive models and an accurate evaluation of large variety of classifiers.
    Downloads: 0 This Week
    Last Update:
    See Project
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 5
    AIMET

    AIMET

    AIMET is a library that provides advanced quantization and compression

    Qualcomm Innovation Center (QuIC) is at the forefront of enabling low-power inference at the edge through its pioneering model-efficiency research. QuIC has a mission to help migrate the ecosystem toward fixed-point inference. With this goal, QuIC presents the AI Model Efficiency Toolkit (AIMET) - a library that provides advanced quantization and compression techniques for trained neural network models. AIMET enables neural networks to run more efficiently on fixed-point AI hardware...
    Downloads: 23 This Week
    Last Update:
    See Project
  • 6
    FlexLLMGen

    FlexLLMGen

    Running large language models on a single GPU

    FlexLLMGen is an open-source inference engine designed to run large language models efficiently on limited hardware resources such as a single GPU. The system focuses on high-throughput generation workloads where large batches of text must be processed quickly, such as large-scale data extraction or document analysis tasks. Instead of requiring expensive multi-GPU systems, the framework uses techniques such as memory offloading, compression, and optimized batching to run large models on...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    tsai

    tsai

    Time series Timeseries Deep Learning Machine Learning Pytorch fastai

    ...New tutorial notebook on how to train your model with larger-than-memory datasets in less time achieving up to 100% GPU usage! See our new tutorial notebook on how to track your experiments with Weights & Biases
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    TextAttack

    TextAttack

    Python framework for adversarial attacks, and data augmentation

    Generating adversarial examples for NLP models. TextAttack is a Python framework for adversarial attacks, data augmentation, and model training in NLP.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    YoloV3 Implemented in TensorFlow 2.0

    YoloV3 Implemented in TensorFlow 2.0

    YoloV3 Implemented in Tensorflow 2.0

    YoloV3 Implemented in TensorFlow 2.0 is built using TensorFlow 2.0. The project provides a modern deep learning implementation of the popular YOLOv3 algorithm, which is widely used for real-time object detection in images and video streams. YOLOv3 works by dividing an image into grid regions and predicting bounding boxes and class probabilities simultaneously, allowing objects to be detected quickly and efficiently. The repository includes training scripts, inference tools, and configuration...
    Downloads: 0 This Week
    Last Update:
    See Project
  • AI-generated apps that pass security review Icon
    AI-generated apps that pass security review

    Stop waiting on engineering. Build production-ready internal tools with AI—on your company data, in your cloud.

    Retool lets you generate dashboards, admin panels, and workflows directly on your data. Type something like “Build me a revenue dashboard on my Stripe data” and get a working app with security, permissions, and compliance built in from day one. Whether on our cloud or self-hosted, create the internal software your team needs without compromising enterprise standards or control.
    Try Retool free
  • 10
    Alphafold2

    Alphafold2

    Unofficial Pytorch implementation / replication of Alphafold2

    ...Once this is replicated, I intend to fold all available amino acid sequences out there in-silico and release it as an academic torrent, to further science. Deepmind has open sourced the official code in Jax, along with the weights! This repository will now be geared towards a straight pytorch translation with some improvements on positional encoding. lhatsk has reported training a modified trunk of this repository, using the same setup as trRosetta, with competitive results. The underlying assumption is that the trunk works on the residue level, and then constitutes to atomic level for the structure module, whether it be SE3 Transformers, E(n)-Transformer, or EGNN doing the refinement.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 11
    YOLOX

    YOLOX

    YOLOX is a high-performance anchor-free YOLO, exceeding yolov3~v5

    ...One more thing worth noting is that you should also implement pull_item and load_anno method for the Mosiac and MixUp augmentations. Except special cases, we always recommend using our COCO pre-trained weights for initializing the model. As YOLOX is an anchor-free detector with only several hyper-parameters, most of the time good results can be obtained with no changes to the models or training settings.
    Downloads: 14 This Week
    Last Update:
    See Project
  • 12
    CodeSearchNet

    CodeSearchNet

    Datasets, tools, and benchmarks for representation learning of code

    CodeSearchNet is a large-scale dataset and research benchmark designed to advance the development of systems that retrieve source code using natural language queries. The project was created through collaboration between GitHub and Microsoft Research and aims to support research on semantic code search and program understanding. The dataset contains millions of pairs of source code functions and corresponding documentation comments extracted from open-source repositories. These pairs allow...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 13
    YOLOv4-large

    YOLOv4-large

    Scaled-YOLOv4: Scaling Cross Stage Partial Network

    YOLOv4-large is an open-source implementation of the Scaled-YOLOv4 object detection architecture, designed to improve both the accuracy and scalability of real-time computer vision models. The project provides a PyTorch implementation of the Scaled-YOLOv4 framework, which extends the original YOLOv4 architecture using Cross Stage Partial (CSP) networks and new scaling techniques.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Minkowski Engine

    Minkowski Engine

    Auto-diff neural network library for high-dimensional sparse tensors

    ...Compressing a neural network to speed up inference and minimize memory footprint has been studied widely. One of the popular techniques for model compression is pruning the weights in convnets, is also known as sparse convolutional networks. Such parameter-space sparsity used for model compression compresses networks that operate on dense tensors and all intermediate activations of these networks are also dense tensors.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Keepsake

    Keepsake

    Version control for machine learning

    Keepsake is a Python library that uploads files and metadata (like hyperparameters) to Amazon S3 or Google Cloud Storage. You can get the data back out using the command-line interface or a notebook.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    PyTorch SimCLR

    PyTorch SimCLR

    PyTorch implementation of SimCLR: A Simple Framework

    ...And most important, their features are known to adapt well to new problems. This is particularly interesting when annotated training data is scarce. In situations like this, we take the models’ pre-trained weights, append a new classifier layer on top of it, and retrain the network. This is called transfer learning, and is one of the most used techniques in CV. Aside from a few tricks when performing fine-tuning (if the case), it has been shown (many times) that if training for a new task, models initialized with pre-trained weights tend to learn faster and be more accurate then training from scratch using random initialization.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    TFLearn

    TFLearn

    Deep learning library featuring a higher-level API for TensorFlow

    ...Powerful helper functions to train any TensorFlow graph, with support of multiple inputs, outputs, and optimizers. Easy and beautiful graph visualization, with details about weights, gradients, activations, and more. Effortless device placement for using multiple CPU/GPU. The high-level API currently supports the most of the recent deep learning models, such as Convolutions, LSTM, BiRNN, BatchNorm, etc.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    Age and Gender Estimation

    Age and Gender Estimation

    Keras implementation of a CNN network for age and gender estimation

    ...In training, the IMDB-WIKI dataset is used. Because the face images in the UTKFace dataset is tightly cropped (there is no margin around the face region), faces should also be cropped in demo.py if weights trained by the UTKFace dataset is used. Please set the margin argument to 0 for tight cropping. You can evaluate a trained model on the APPA-REAL (validation) dataset. We pose the age regression problem as a deep classification problem followed by a softmax expected value refinement and show improvements over direct regression training of CNNs. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 19
    StellarGraph

    StellarGraph

    Machine Learning on Graphs

    StellarGraph is a Python library for machine learning on graphs and networks. The StellarGraph library offers state-of-the-art algorithms for graph machine learning, making it easy to discover patterns and answer questions about graph-structured data. It can solve many machine learning tasks. Graph-structured data represent entities as nodes (or vertices) and relationships between them as edges (or links), and can include data associated with either as attributes. For example, a graph can...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 20
    TensorNets

    TensorNets

    High level network definitions with pre-trained weights in TensorFlow

    ...Also, it is easy to deploy and expand a collection of pre-processing and pre-trained weights. Readability. With recent TensorFlow APIs, more factoring and less indenting can be possible. For example, all the inception variants are implemented as about 500 lines of code in TensorNets while 2000+ lines in official TensorFlow models. Reproducibility. You can always reproduce the original results with simple APIs including feature extractions.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    textgenrnn

    textgenrnn

    Easily train your own text-generating neural network

    With textgenrnn you can easily train your own text-generating neural network of any size and complexity on any text dataset with a few lines of code. A modern neural network architecture that utilizes new techniques as attention-weighting and skip-embedding to accelerate training and improve model quality. Train on and generate text at either the character-level or word-level. Configure RNN size, the number of RNN layers, and whether to use bidirectional RNNs. Train on any generic input text...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    Image Super-Resolution (ISR)

    Image Super-Resolution (ISR)

    Super-scale your images and run experiments with Residual Dense

    ...When training your own model, start with only PSNR loss (50+ epochs, depending on the dataset) and only then introduce GANS and feature loss. This can be controlled by the loss weights argument. The weights used to produce these images are available directly when creating the model object. ISR is compatible with Python 3.6 and is distributed under the Apache 2.0 license.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 23
    Torchreid

    Torchreid

    Deep learning person re-identification in PyTorch

    ...The folder "configs/" contains some predefined configs which you can use as a starting point. The code will automatically (download and) load the ImageNet pretrained weights. After the training is done, the model will be saved as "log/osnet_x1_0_market1501_softmax_cosinelr/model.pth.tar-250". Under the same folder, you can find the tensorboard file. Different from the same-domain setting, here we replace random_erase with color_jitter. This can improve the generalization performance on the unseen target dataset.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    ELI5

    ELI5

    A library for debugging/inspecting machine learning classifiers

    ...The project focuses on improving model transparency by providing tools that visualize feature importance and prediction reasoning. It supports several popular machine learning frameworks including scikit-learn, XGBoost, LightGBM, CatBoost, and Keras. The library allows users to inspect model weights, analyze decision trees, and compute permutation feature importance for black-box models. It also provides specialized tools such as TextExplainer, which can highlight important words in text classification tasks to explain why a model produced a particular prediction. Additionally, the library integrates explanation algorithms such as LIME to interpret predictions from arbitrary machine learning models.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    FID score for PyTorch

    FID score for PyTorch

    Compute FID scores with PyTorch

    ...It was shown to correlate well with human judgement of visual quality and is most often used to evaluate the quality of samples of Generative Adversarial Networks. FID is calculated by computing the Fréchet distance between two Gaussians fitted to feature representations of the Inception network. The weights and the model are exactly the same as in the official Tensorflow implementation, and were tested to give very similar results (e.g. .08 absolute error and 0.0009 relative error on LSUN, using ProGAN generated images). However, due to differences in the image interpolation implementation and library backends, FID results still differ slightly from the original implementation. ...
    Downloads: 6 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • Next
MongoDB Logo MongoDB