Search Results for "mime decoder encoder" - Page 2

Showing 90 open source projects for "mime decoder encoder"

View related business solutions
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 1
    CPT

    CPT

    CPT: A Pre-Trained Unbalanced Transformer

    A Pre-Trained Unbalanced Transformer for Both Chinese Language Understanding and Generation. We replace the old BERT vocabulary with a larger one of size 51271 built from the training data, in which we 1) add missing 6800+ Chinese characters (most of them are traditional Chinese characters); 2) remove redundant tokens (e.g. Chinese character tokens with ## prefix); 3) add some English tokens to reduce OOV. Position Embeddings We extend the max_position_embeddings from 512 to 1024. We...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 2
    Karlo

    Karlo

    Text-conditional image generation model based on OpenAI's unCLIP

    Karlo is a text-conditional image generation model based on OpenAI's unCLIP architecture with the improvement over the standard super-resolution model from 64px to 256px, recovering high-frequency details only in the small number of denoising steps. We train all components from scratch on 115M image-text pairs including COYO-100M, CC3M, and CC12M. In the case of Prior and Decoder, we use ViT-L/14 provided by OpenAI’s CLIP repository. Unlike the original implementation of unCLIP, we replace the trainable transformer in the decoder into the text encoder in ViT-L/14 for efficiency. In the case of the SR module, we first train the model using the DDPM objective in 1M steps, followed by additional 234K steps to fine-tune the additional component.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    LightSeq

    LightSeq

    A High Performance Library for Sequence Processing and Generation

    Lightseq is a high-performance library focused on efficient inference and training for deep learning models, especially large language models (LLMs) and transformer-based architectures. Its goal is to optimize both memory usage and computational throughput, enabling faster training or inference on limited hardware while maintaining model quality. Lightseq provides optimized CUDA kernels, quantization strategies, and runtime optimizations tailored for transformer operations — which often are...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    EnCodec

    EnCodec

    State-of-the-art deep learning based audio codec

    Encodec is a neural audio codec developed by Meta for high-fidelity, low-bitrate audio compression using end-to-end deep learning. Unlike traditional codecs (like MP3 or Opus), Encodec uses a learned quantizer and decoder to reconstruct complex waveforms with remarkable accuracy at bitrates as low as 1.5 kbps. It employs a convolutional encoderdecoder architecture trained with perceptual loss functions that optimize for human auditory quality rather than raw waveform distance. The model can operate in real time and supports variable bandwidths, bitrates, and multi-band audio. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Application Monitoring That Won't Slow Your App Down Icon
    Application Monitoring That Won't Slow Your App Down

    AppSignal's Rust-based agent is lightweight and stable. Already running in thousands of production apps.

    Full APM with errors, performance, logs, and uptime monitoring. 99.999% uptime SLA on the platform itself.
    Start Free
  • 5
    MAE (Masked Autoencoders)

    MAE (Masked Autoencoders)

    PyTorch implementation of MAE

    ...It trains a Vision Transformer (ViT) by randomly masking a high percentage of image patches (typically 75%) and reconstructing the missing content from the remaining visible patches. This forces the model to learn semantic structure and global context without supervision. The encoder processes only the visible patches, while a lightweight decoder reconstructs the full image—making pretraining computationally efficient. After pretraining, the encoder serves as a powerful backbone for downstream tasks like image classification, segmentation, and detection, achieving top performance with minimal fine-tuning. The repository provides pretrained models, fine-tuning scripts, evaluation protocols, and visualization tools for reconstruction quality and learned features.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Deep learning time series forecasting

    Deep learning time series forecasting

    Deep learning PyTorch library for time series forecasting

    ...Historically, this repository provided open-source benchmarks and codes for flash flood and river flow forecasting. Full transformer (SimpleTransformer in model_dict): The full original transformer with all 8 encoder and decoder blocks. Requires passing the target in at inference.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Reformer PyTorch

    Reformer PyTorch

    Reformer, the efficient Transformer, in Pytorch

    This is a Pytorch implementation of Reformer. It includes LSH attention, reversible network, and chunking. It has been validated with an auto-regressive task (enwik8).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    fairseq-lua

    fairseq-lua

    Facebook AI Research Sequence-to-Sequence Toolkit

    fairseq-lua is the original Lua/Torch7 version of Facebook AI Research’s sequence modeling toolkit, designed for neural machine translation (NMT) and sequence generation. It introduced early attention-based architectures and training pipelines that later evolved into the modern PyTorch-based fairseq. The framework implements sequence-to-sequence models with attention, beam search decoding, and distributed training, providing a research platform for exploring translation, summarization, and...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9

    file-splitter-rejoiner

    file splitter and rejoiner

    /* * * Freeware * Open Source * 2 tools in one application * using .Net 4.8 * (1) Simple files splitter and rejoiner tool using memory buffer * (2) Simple files base64 encoder and decoder using random sized Stream GB/TB+ data sizes * A good tool for an essentials inventory * Just when required. * Simple precise short and straightforward coding * Tested bugs free and perfect when I developed and released it. * * Developer: Tushar Jain * Release Time: 09:33 PM * Release Date: Friday, 23 April 2021 * * */
    Downloads: 0 This Week
    Last Update:
    See Project
  • Go From AI Idea to AI App Fast Icon
    Go From AI Idea to AI App Fast

    One platform to build, fine-tune, and deploy ML models. No MLOps team required.

    Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
    Try Free
  • 10
    AliceMind

    AliceMind

    ALIbaba's Collection of Encoder-decoders from MinD

    This repository provides pre-trained encoder-decoder models and its related optimization techniques developed by Alibaba's MinD (Machine IntelligeNce of Damo) Lab. Pre-trained models for natural language understanding (NLU). We extend BERT to a new model, StructBERT, by incorporating language structures into pre-training. Specifically, we pre-train StructBERT with two auxiliary tasks to make the most of the sequential order of words and sentences, which leverage language structures at the word and sentence levels, respectively. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Denoiser

    Denoiser

    Real Time Speech Enhancement in the Waveform Domain (Interspeech 2020)

    Denoiser is a real-time speech enhancement model operating directly on raw waveforms, designed to clean noisy audio while running efficiently on CPU. It uses a causal encoder-decoder architecture with skip connections, optimized with losses defined both in the time domain and frequency domain to better suppress noise while preserving speech. Unlike models that operate on spectrograms alone, this design enables lower latency and coherent waveform output. The implementation includes data augmentation techniques applied to the raw waveforms (e.g. noise mixing, reverberation) to improve model robustness and generalization to diverse noise types. ...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 12
    Abdal Base64 Encoder Decoder

    Abdal Base64 Encoder Decoder

    Base64 Encoder Decoder Tools

    Abdal Base64 Encoder Decoder tool can perform encryption and decryption process without any restrictions and completely free of charge, but be careful that this tool should not be used for encrypting very sensitive information.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    ALAE

    ALAE

    Adversarial Latent Autoencoders

    ...The project implements the architecture introduced in the CVPR research paper on Adversarial Latent Autoencoders, which focuses on improving generative modeling by learning latent representations aligned with adversarial training objectives. Unlike traditional GANs that directly generate images from random noise, ALAE uses an encoder-decoder architecture that maps images into a structured latent space and then reconstructs them through adversarial training. This design allows the model to learn interpretable latent representations that can be manipulated to control generated image attributes.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    clzw

    clzw

    Implementation of LZW compression algorithm in C

    Simple, fast implementation of LZW (Lempel–Ziv–Welch) data compression algorithm in C. - Console encoder/decoder tools - OS independent - Could be used in embedded projects - Works with raw code-stream LZW features: - Hardcoded dictionary size - Variable code size - Code search is performed by hash table and embedded in dictionary linked lists (encoder) - No dynamic memory allocation
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15

    Fox Encoder

    Multipurpose Encoder/decoder

    Encode / Decode (from and to) text, hex, base64. Also supports hash function including MD4, MD5, SHA1, SHA2, SHA3, Keccak including most lengths (224, 256, 384 and 512),
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    DETR

    DETR

    End-to-end object detection with transformers

    ...Unlike traditional computer vision techniques, DETR approaches object detection as a direct set prediction problem. It consists of a set-based global loss, which forces unique predictions via bipartite matching, and a Transformer encoder-decoder architecture. Given a fixed small set of learned object queries, DETR reasons about the relations of the objects and the global image context to directly output the final set of predictions in parallel. Due to this parallel nature, DETR is very fast and efficient.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Universal Encoder Decoder - AyaN Softwar

    Universal Encoder Decoder - AyaN Softwar

    84 Type Encoding/Decoding Options And Full Offline - AyaN Software

    In the era of digital communication and data security and computer management the character arranging encoding and decoding system is doing its best. You can encode and decode data easily with the online tools but this the software Universal Encoder Decoder can do all types of encoding and decoding as fast as light. Some most advance feature of this encoding and decoding is given below , which make this software different from others : This software - first of all easy to install and easy to use, just one click and it will start. There is no difficulties of installing process and system agreement. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 18
    Texar

    Texar

    Toolkit for Machine Learning, Natural Language Processing

    Texar is a toolkit aiming to support a broad set of machine learning, especially natural language processing and text generation tasks. Texar provides a library of easy-to-use ML modules and functionalities for composing whatever models and algorithms. The tool is designed for both researchers and practitioners for fast prototyping and experimentation. Texar was originally developed and is actively contributed by Petuum and CMU in collaboration with other institutes. A mirror of this...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    GoJay

    GoJay

    high performance JSON encoder/decoder with stream API for Golang

    GoJay is a performant JSON encoder/decoder for Golang (currently the most performant, see benchmarks). It has a simple API and doesn't use reflection. It relies on small interfaces to decode/encode structures and slices. Gojay also comes with powerful stream decoding features and an even faster Unsafe API. There is also a code generation tool to make usage easier and faster.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    CakeChat

    CakeChat

    CakeChat: Emotional Generative Dialog System

    ...The code is flexible and allows to condition model's responses by an arbitrary categorical variable. For example, you can train your own persona-based neural conversational model or create an emotional chatting machine. Hierarchical Recurrent Encoder-Decoder (HRED) architecture for handling deep dialog context. Multilayer RNN with GRU cells. The first layer of the utterance-level encoder is always bidirectional. By default, CuDNNGRU implementation is used for ~25% acceleration during inference. Thought vector is fed into decoder on each decoding step. Decoder can be conditioned on any categorical label, for example, emotion label or persona id. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    SYPPS

    SYPPS

    small yet powerful php shell

    SYPPS - small yet powerful php shell is another PHP shell for pentesting
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    OpenSeq2Seq

    OpenSeq2Seq

    Toolkit for efficient experimentation with Speech Recognition

    OpenSeq2Seq is a TensorFlow-based toolkit for efficient experimentation with sequence-to-sequence models across speech and NLP tasks. Its core goal is to give researchers a flexible, modular framework for building and training encoderdecoder architectures while fully leveraging distributed and mixed-precision training. The toolkit includes ready-made models for neural machine translation, automatic speech recognition, speech synthesis, language modeling, and additional NLP tasks such as sentiment analysis. It supports multi-GPU and multi-node data-parallel training, and integrates with Horovod to scale out across large GPU clusters. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    he

    he

    A robust HTML entity encoder/decoder written in JavaScript

    he is a JavaScript library that provides robust HTML entity encoding and decoding, with full Unicode support. It supports all standardized named character references (e.g., ©, —), handles numeric and hex entities, and deals properly with astral Unicode symbols (i.e., code points outside the BMP). The library is designed so that he.decode(input) will safely convert HTML-entity encoded strings into proper Unicode text, and he.encode(text, options) will encode non-ASCII or special...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    bin64ed

    bin64ed

    Base64 encode (or decode) files of any type with this lightweight tool

    bin64ed is a binary base64 encoder/decoder that allows you to encode/decode binary files (such as images, pdfs, etc) to and from base64.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 25
    WOFF2

    WOFF2

    This document documents how to run the compression reference code

    ...The repository includes a compact C/C++ library and small command-line tools so you can convert existing TTF/OTF files to WOFF2 and back for testing or build pipelines. Its encoder applies deterministic, spec-compliant transformations that maximize compressibility without altering rendering results, making it safe for production web delivery. The decoder is just as strict, validating headers and table checksums to guard against malformed inputs. Because WOFF2 is now ubiquitous across browsers and CDNs, this repo often serves as the canonical baseline for tooling.
    Downloads: 12 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB