Search Results for "sacd decoder mac" - Page 3

Showing 242 open source projects for "sacd decoder mac"

View related business solutions
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • Go From AI Idea to AI App Fast Icon
    Go From AI Idea to AI App Fast

    One platform to build, fine-tune, and deploy ML models. No MLOps team required.

    Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
    Try Free
  • 1
    OpenNMT-tf

    OpenNMT-tf

    Neural machine translation and sequence learning using TensorFlow

    OpenNMT is an open-source ecosystem for neural machine translation and neural sequence learning. OpenNMT-tf is a general-purpose sequence learning toolkit using TensorFlow 2. While neural machine translation is the main target task, it has been designed to more generally support sequence-to-sequence mapping, sequence tagging, sequence classification, language modeling. Models are described with code to allow training custom architectures and overriding default behavior. For example, the...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2

    PNG_IO

    An Ada 2012 library for reading and writing PNG image files

    PNG_IO is an Ada2012 coder/decoder for Portable Network Graphics, allowing Ada2012 programs to read and write PNG files. It is portable across machines and operating systems. It depends on Zlib and Zlib_Ada to perform lossless compression and decompression. Versions prior to v5 were in Ada95.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    FasterTransformer

    FasterTransformer

    Transformer related optimization, including BERT, GPT

    FasterTransformer is a high-performance inference library designed to accelerate transformer-based models such as BERT, GPT, and T5 on NVIDIA GPUs. It provides optimized implementations of transformer encoder and decoder layers using CUDA, cuBLAS, and custom kernels to maximize throughput and minimize latency. The library supports multiple deep learning frameworks, including TensorFlow, PyTorch, and Triton, allowing developers to integrate it into existing pipelines without major changes. It...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    NÜWA - Pytorch

    NÜWA - Pytorch

    Implementation of NÜWA, attention network for text to video synthesis

    Implementation of NÜWA, state of the art attention network for text-to-video synthesis, in Pytorch. It also contains an extension into video and audio generation, using a dual decoder approach. It seems as though a diffusion-based method has taken the new throne for SOTA. However, I will continue on with NUWA, extending it to use multi-headed codes + hierarchical causal transformer. I think that direction is untapped for improving on this line of work. In the paper, they also present a way...
    Downloads: 0 This Week
    Last Update:
    See Project
  • AI-generated apps that pass security review Icon
    AI-generated apps that pass security review

    Stop waiting on engineering. Build production-ready internal tools with AI—on your company data, in your cloud.

    Retool lets you generate dashboards, admin panels, and workflows directly on your data. Type something like “Build me a revenue dashboard on my Stripe data” and get a working app with security, permissions, and compliance built in from day one. Whether on our cloud or self-hosted, create the internal software your team needs without compromising enterprise standards or control.
    Try Retool free
  • 5
    Lyra

    Lyra

    A Very Low-Bitrate Codec for Speech Compression

    lyra is a neural audio codec designed to deliver intelligible, natural-sounding speech at extremely low bitrates, making real-time communication viable on constrained networks. It replaces hand-engineered codecs with learned models that capture speech characteristics more efficiently and reconstruct waveforms with a neural vocoder. The system targets mobile-class hardware, balancing latency and quality so it can run in real-time on phones. Its architecture is resilient to packet loss and...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Karlo

    Karlo

    Text-conditional image generation model based on OpenAI's unCLIP

    Karlo is a text-conditional image generation model based on OpenAI's unCLIP architecture with the improvement over the standard super-resolution model from 64px to 256px, recovering high-frequency details only in the small number of denoising steps. We train all components from scratch on 115M image-text pairs including COYO-100M, CC3M, and CC12M. In the case of Prior and Decoder, we use ViT-L/14 provided by OpenAI’s CLIP repository. Unlike the original implementation of unCLIP, we...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    CPT

    CPT

    CPT: A Pre-Trained Unbalanced Transformer

    A Pre-Trained Unbalanced Transformer for Both Chinese Language Understanding and Generation. We replace the old BERT vocabulary with a larger one of size 51271 built from the training data, in which we 1) add missing 6800+ Chinese characters (most of them are traditional Chinese characters); 2) remove redundant tokens (e.g. Chinese character tokens with ## prefix); 3) add some English tokens to reduce OOV. Position Embeddings We extend the max_position_embeddings from 512 to 1024. We...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 8
    LightSeq

    LightSeq

    A High Performance Library for Sequence Processing and Generation

    Lightseq is a high-performance library focused on efficient inference and training for deep learning models, especially large language models (LLMs) and transformer-based architectures. Its goal is to optimize both memory usage and computational throughput, enabling faster training or inference on limited hardware while maintaining model quality. Lightseq provides optimized CUDA kernels, quantization strategies, and runtime optimizations tailored for transformer operations — which often are...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    EnCodec

    EnCodec

    State-of-the-art deep learning based audio codec

    Encodec is a neural audio codec developed by Meta for high-fidelity, low-bitrate audio compression using end-to-end deep learning. Unlike traditional codecs (like MP3 or Opus), Encodec uses a learned quantizer and decoder to reconstruct complex waveforms with remarkable accuracy at bitrates as low as 1.5 kbps. It employs a convolutional encoder–decoder architecture trained with perceptual loss functions that optimize for human auditory quality rather than raw waveform distance. The model can...
    Downloads: 0 This Week
    Last Update:
    See Project
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 10
    project_bwt
    This is a research project to possible improve compression using LZ77, LZSS or bwt, bwt is used on the bzip2 compression library, the v28 from 04/dec/2022 is the first version to almost beat lzma, indeed it is a tweak of lzma, the decoder for v28 will follow in a few days
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    DiffSinger

    DiffSinger

    Singing Voice Synthesis via Shallow Diffusion Mechanism

    DiffSinger is an open-source PyTorch implementation of a diffusion-based acoustic model for singing-voice synthesis (SVS) and also text-to-speech (TTS) in a related variant. The core idea is to view generation of a sung voice (mel-spectrogram) as a diffusion process: starting from noise, the model iteratively “denoises” while being conditioned on a music score (lyrics, pitch, musical timing). This avoids some of the typical problems of prior SVS models — like over-smoothing or unstable GAN...
    Downloads: 41 This Week
    Last Update:
    See Project
  • 12
    LaMDA-pytorch

    LaMDA-pytorch

    Open-source pre-training implementation of Google's LaMDA in PyTorch

    Open-source pre-training implementation of Google's LaMDA research paper in PyTorch. The totally not sentient AI. This repository will cover the 2B parameter implementation of the pre-training architecture as that is likely what most can afford to train. You can review Google's latest blog post from 2022 which details LaMDA here. You can also view their previous blog post from 2021 on the model.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    JSON Parser

    JSON Parser

    JSON Parser written in C that outputs JSON into D-List list objects

    JSON parser is a highly efficient JSON data decoder written in C, that parses JSON data into a form that is very easy to use from within C code. You only need to include the json-parser.h file in your source code. If your project does not already incorporate D-List, you will also need to include dlist.h The JSON parser package also includes a JSON validator that does not parse or create resources, but allows the caller to validate JSON data from unknown sources, prior to any parsing...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 14
    Mask2Former

    Mask2Former

    Code release for "Masked-attention Mask Transformer

    Mask2Former is a unified segmentation architecture that handles semantic, instance, and panoptic segmentation with one model and one training recipe. Its core idea is to cast segmentation as mask classification: a transformer decoder predicts a set of mask queries, each with an associated class score, eliminating the need for task-specific heads. A pixel decoder fuses multi-scale features and feeds masked attention in the transformer so each query focuses computation on its current spatial...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Deep learning time series forecasting

    Deep learning time series forecasting

    Deep learning PyTorch library for time series forecasting

    Example image Flow Forecast (FF) is an open-source deep learning for time series forecasting framework. It provides all the latest state-of-the-art models (transformers, attention models, GRUs) and cutting-edge concepts with easy-to-understand interpretability metrics, cloud provider integration, and model serving capabilities. Flow Forecast was the first time series framework to feature support for transformer-based models and remains the only true end-to-end deep learning for time series...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    MAE (Masked Autoencoders)

    MAE (Masked Autoencoders)

    PyTorch implementation of MAE

    MAE (Masked Autoencoders) is a self-supervised learning framework for visual representation learning using masked image modeling. It trains a Vision Transformer (ViT) by randomly masking a high percentage of image patches (typically 75%) and reconstructing the missing content from the remaining visible patches. This forces the model to learn semantic structure and global context without supervision. The encoder processes only the visible patches, while a lightweight decoder reconstructs the...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    pcdtojpeg converts PCD (Kodak Photo CD) image files into high quality JPEG images. Conversions are at the maximum available PCD resolution, and are fully color managed - no blown highlights or color casts. A C++ PCD decoder library is also available.
    Leader badge
    Downloads: 33 This Week
    Last Update:
    See Project
  • 18
    Logisim Italian Fork

    Logisim Italian Fork

    in progress...

    Github: https://github.com/LogisimIt/Logisim
    Downloads: 47 This Week
    Last Update:
    See Project
  • 19
    Reformer PyTorch

    Reformer PyTorch

    Reformer, the efficient Transformer, in Pytorch

    This is a Pytorch implementation of Reformer. It includes LSH attention, reversible network, and chunking. It has been validated with an auto-regressive task (enwik8).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    fairseq-lua

    fairseq-lua

    Facebook AI Research Sequence-to-Sequence Toolkit

    fairseq-lua is the original Lua/Torch7 version of Facebook AI Research’s sequence modeling toolkit, designed for neural machine translation (NMT) and sequence generation. It introduced early attention-based architectures and training pipelines that later evolved into the modern PyTorch-based fairseq. The framework implements sequence-to-sequence models with attention, beam search decoding, and distributed training, providing a research platform for exploring translation, summarization, and...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21

    file-splitter-rejoiner

    file splitter and rejoiner

    /* * * Freeware * Open Source * 2 tools in one application * using .Net 4.8 * (1) Simple files splitter and rejoiner tool using memory buffer * (2) Simple files base64 encoder and decoder using random sized Stream GB/TB+ data sizes * A good tool for an essentials inventory * Just when required. * Simple precise short and straightforward coding * Tested bugs free and perfect when I developed and released it. * * Developer: Tushar Jain * Release Time: 09:33 PM *...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    AliceMind

    AliceMind

    ALIbaba's Collection of Encoder-decoders from MinD

    This repository provides pre-trained encoder-decoder models and its related optimization techniques developed by Alibaba's MinD (Machine IntelligeNce of Damo) Lab. Pre-trained models for natural language understanding (NLU). We extend BERT to a new model, StructBERT, by incorporating language structures into pre-training. Specifically, we pre-train StructBERT with two auxiliary tasks to make the most of the sequential order of words and sentences, which leverage language structures at the...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Denoiser

    Denoiser

    Real Time Speech Enhancement in the Waveform Domain (Interspeech 2020)

    Denoiser is a real-time speech enhancement model operating directly on raw waveforms, designed to clean noisy audio while running efficiently on CPU. It uses a causal encoder-decoder architecture with skip connections, optimized with losses defined both in the time domain and frequency domain to better suppress noise while preserving speech. Unlike models that operate on spectrograms alone, this design enables lower latency and coherent waveform output. The implementation includes data...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    Abdal Base64 Encoder Decoder

    Abdal Base64 Encoder Decoder

    Base64 Encoder Decoder Tools

    Abdal Base64 Encoder Decoder tool can perform encryption and decryption process without any restrictions and completely free of charge, but be careful that this tool should not be used for encrypting very sensitive information.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 25
    ALAE

    ALAE

    Adversarial Latent Autoencoders

    ALAE (Adversarial Latent Autoencoders) is a deep learning research implementation that combines autoencoders with generative adversarial networks to produce high-quality image synthesis models. The project implements the architecture introduced in the CVPR research paper on Adversarial Latent Autoencoders, which focuses on improving generative modeling by learning latent representations aligned with adversarial training objectives. Unlike traditional GANs that directly generate images from...
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB