Showing 206 open source projects for "input-leap"

View related business solutions
  • Our Free Plans just got better! | Auth0 Icon
    Our Free Plans just got better! | Auth0

    With up to 25k MAUs and unlimited Okta connections, our Free Plan lets you focus on what you do best—building great apps.

    You asked, we delivered! Auth0 is excited to expand our Free and Paid plans to include more options so you can focus on building, deploying, and scaling applications without having to worry about your security. Auth0 now, thank yourself later.
    Try free now
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    Access competitive interest rates on your digital assets.

    Generate interest, borrow against your crypto, and trade a range of cryptocurrencies — all in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 1
    OpenNMT-tf

    OpenNMT-tf

    Neural machine translation and sequence learning using TensorFlow

    ...Models are described with code to allow training custom architectures and overriding default behavior. For example, the following instance defines a sequence-to-sequence model with 2 concatenated input features, a self-attentional encoder, and an attentional RNN decoder sharing its input and output embeddings. Sequence to sequence models can be trained with guided alignment and alignment information are returned as part of the translation API.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    Free-Auto-GPT

    Free-Auto-GPT

    Free AutoGPT enables autonomous AI tasks without paid APIs

    ...It allows users to run an AutoGPT-style system without relying on paid OpenAI APIs, making it more accessible for experimentation and personal use. Free Auto GPT can take a goal, break it into smaller steps, and execute actions in a loop to achieve results with minimal human input. Designed for ease of use, the project focuses on removing cost barriers while still demonstrating how autonomous agents function. It is suitable for developers, learners, and hobbyists who want to explore AI-driven automation without subscription requirements. It provides a lightweight implementation that highlights core AutoGPT concepts such as task decomposition, iteration, and independent execution in a simplified environment.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 3
    TensorFlow Documentation

    TensorFlow Documentation

    TensorFlow documentation

    An end-to-end platform for machine learning. TensorFlow makes it easy to create ML models that can run in any environment. Learn how to use the intuitive APIs through interactive code samples.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Simple LLM Finetuner

    Simple LLM Finetuner

    Simple UI for LLM Model Finetuning

    ...It allows users to customize pre-trained models using relatively small datasets and modest hardware, making it feasible to experiment with LLM training even on consumer-grade GPUs or cloud environments like Google Colab. The tool includes a web-based interface where users can input datasets, configure training parameters, and run fine-tuning jobs without deep knowledge of machine learning pipelines. It leverages libraries such as Hugging Face PEFT to enable efficient adaptation of models by modifying only a subset of parameters, significantly reducing computational requirements. In addition to training, the platform provides inference capabilities so users can immediately test and evaluate their fine-tuned models within the same environment.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 5
    Hyperformer

    Hyperformer

    Hypergraph Transformer for Skeleton-based Action Recognition

    ...More recently, a limitation of GCNs is identified, i.e., the topology is fixed after training. To relax such a restriction, Self-Attention (SA) mechanism has been adopted to make the topology of GCNs adaptive to the input, resulting in the state-of-the-art hybrid models. Concurrently, attempts with plain Transformers have also been made, but they still lag behind state-of-the-art GCN-based methods due to the lack of structural prior.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    ArtLine

    ArtLine

    Deep learning tool that converts portrait photos into line art

    ArtLine is a deep learning-based project focused on generating high-quality line art portraits from input images. It leverages neural network techniques built on top of the fastai library and PyTorch to transform photographic portraits into stylized line drawings. ArtLine is trained using datasets such as APDrawing and anime sketch colorization pairs to better understand facial structures and artistic line representation. An extended version integrates ControlNet, allowing users to guide the output style through textual instructions alongside the input image. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 7
    CodeContests

    CodeContests

    Large dataset of coding contests designed for AI and ML model training

    ...CodeContests aggregates problems and human-written solutions from multiple programming competition platforms, including AtCoder, Codeforces, CodeChef, Aizu, and HackerEarth. Each problem includes structured metadata, problem descriptions, paired input/output test cases, and multiple correct and incorrect solutions in various programming languages. The dataset is distributed in Riegeli format using Protocol Buffers, with separate training, validation, and test splits for reproducible machine learning experiments.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 8
    CPT

    CPT

    CPT: A Pre-Trained Unbalanced Transformer

    A Pre-Trained Unbalanced Transformer for Both Chinese Language Understanding and Generation. We replace the old BERT vocabulary with a larger one of size 51271 built from the training data, in which we 1) add missing 6800+ Chinese characters (most of them are traditional Chinese characters); 2) remove redundant tokens (e.g. Chinese character tokens with ## prefix); 3) add some English tokens to reduce OOV. Position Embeddings We extend the max_position_embeddings from 512 to 1024. We...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 9
    Emb-GAM

    Emb-GAM

    An interpretable and efficient predictor using pre-trained models

    ...In contrast, generalized additive models (GAMs) can maintain interpretability but often suffer from poor prediction performance due to their inability to effectively capture feature interactions. In this work, we aim to bridge this gap by using pre-trained neural language models to extract embeddings for each input before learning a linear model in the embedding space. The final model (which we call Emb-GAM) is a transparent, linear function of its input features and feature interactions. Leveraging the language model allows Emb-GAM to learn far fewer linear coefficients, model larger interactions, and generalize well to novel inputs. Across a variety of natural-language-processing datasets, Emb-GAM achieves strong prediction performance without sacrificing interpretability.
    Downloads: 0 This Week
    Last Update:
    See Project
  • AI-powered service management for IT and enterprise teams Icon
    AI-powered service management for IT and enterprise teams

    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity. Maximize operational efficiency with refreshingly simple, AI-powered Freshservice.
    Try it Free
  • 10
    DiffSinger

    DiffSinger

    Singing Voice Synthesis via Shallow Diffusion Mechanism

    DiffSinger is an open-source PyTorch implementation of a diffusion-based acoustic model for singing-voice synthesis (SVS) and also text-to-speech (TTS) in a related variant. The core idea is to view generation of a sung voice (mel-spectrogram) as a diffusion process: starting from noise, the model iteratively “denoises” while being conditioned on a music score (lyrics, pitch, musical timing). This avoids some of the typical problems of prior SVS models — like over-smoothing or unstable GAN...
    Downloads: 29 This Week
    Last Update:
    See Project
  • 11
    Auto-PyTorch

    Auto-PyTorch

    Automatic architecture search and hyperparameter optimization

    While early AutoML frameworks focused on optimizing traditional ML pipelines and their hyperparameters, another trend in AutoML is to focus on neural architecture search. To bring the best of these two worlds together, we developed Auto-PyTorch, which jointly and robustly optimizes the network architecture and the training hyperparameters to enable fully automated deep learning (AutoDL). Auto-PyTorch is mainly developed to support tabular data (classification, regression) and time series...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Talking Head Anime from a Single Image

    Talking Head Anime from a Single Image

    Demo for the "Talking Head Anime from a Single Image"

    Talking Head Anime from a Single Image is a machine learning project that demonstrates how neural networks can animate anime characters using only a single input image. The system generates animated facial expressions and movements by applying pose transformations to a static image of an anime character. The underlying model uses deep learning techniques to predict how different facial features and body parts should move based on pose parameters or input signals. This allows the software to create realistic animated frames while preserving the identity and appearance of the original character. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    nlpaug

    nlpaug

    Data augmentation for NLP

    This Python library helps you with augmenting nlp for your machine learning projects. Visit this introduction to understand Data Augmentation in NLP. Augmenter is the basic element of augmentation while Flow is a pipeline to orchestra multi augmenters together.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Disco Diffusion

    Disco Diffusion

    Notebooks, models and techniques for the generation of AI Art

    A frankensteinian amalgamation of notebooks, models, and techniques for the generation of AI art and animations. This project uses a special conversion tool to convert the Python files into notebooks for easier development. What this means is you do not have to touch the notebook directly to make changes to it. The tool being used is called Colab-Convert. Initial QoL improvements added, including user-friendly UI, settings+prompt saving, and improved google drive folder organization. Now...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    AI Atelier

    AI Atelier

    Based on the Disco Diffusion, version of the AI art creation software

    ...When a modified version is used to provide a service over a network, the complete source code of the modified version must be made available. Create 2D and 3D animations and not only still frames (from Disco Diffusion v5 and VQGAN Animations). Input audio and images for generation instead of just text. Simplify tool setup process on colab, and enable ‘one-click’ sharing of the generated link to other users. Experiment with the possibilities for multi-user access to the same link.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 16
    PromptSource

    PromptSource

    Toolkit for creating, sharing and using natural language prompts

    ...This emphasizes the need for new tools to create, share and use natural language prompts. Prompts are functions that map an example from a dataset to a natural language input and target output. PromptSource contains a growing collection of prompts (which we call P3: Public Pool of Prompts). As of January 20, 2022, there are ~2'000 English prompts for 170+ English datasets in P3.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 17
    OpenPrompt

    OpenPrompt

    An Open-Source Framework for Prompt-Learning

    ...The template is one of the most important modules in prompt learning, which wraps the original input with textual or soft-encoding sequence. Use the implementations of current prompt-learning approaches.* We have implemented various of prompting methods, including templating, verbalizing and optimization strategies under a unified standard. You can easily call and understand these methods.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    GLIDE (Text2Im)

    GLIDE (Text2Im)

    GLIDE: a diffusion-based text-conditional image synthesis model

    ...The repository provides both model code and pretrained checkpoints, making it possible for researchers and developers to experiment with text-to-image synthesis. GLIDE includes advanced techniques such as classifier-free guidance, which improves the quality and alignment of generated images with the input text. The project also offers sampling scripts and utilities for exploring how diffusion models can be applied to multimodal tasks. As one of the early diffusion-based text-to-image systems, glide-text2im laid important groundwork for later advances in generative AI research.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 19
    TensorFlow Backend for ONNX

    TensorFlow Backend for ONNX

    Tensorflow Backend for ONNX

    ...ONNX is supported by a community of partners who have implemented it in many frameworks and tools. TensorFlow Backend for ONNX makes it possible to use ONNX models as input for TensorFlow. The ONNX model is first converted to a TensorFlow model and then delegated for execution on TensorFlow to produce the output. This is one of the two TensorFlow converter projects which serve different purposes in the ONNX community. ONNX-TF requires ONNX (Open Neural Network Exchange) as an external dependency, for any issues related to ONNX installation, we refer our users to ONNX project repository for documentation and help. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    Interactive Deep Colorization

    Interactive Deep Colorization

    Deep learning software for colorizing black and white images

    Interactive Deep Colorization is a software project for colorizing black-and-white (grayscale) images using deep learning, allowing users to add a few hints (e.g. scribbles) and get a plausible, fully colorized output. The idea is to merge automatic colorization (via neural networks) with optional user guidance — so if the automatic model’s guess isn’t quite right, the user can nudge colors via hints to steer the result, achieving more controlled, satisfying outputs. The project includes...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    VQGAN-CLIP web app

    VQGAN-CLIP web app

    Local image generation using VQGAN-CLIP or CLIP guided diffusion

    ...Searching the r/deepdream subreddit for VQGAN-CLIP yields quite a number of results. Basically, VQGAN can generate pretty high-fidelity images, while CLIP can produce relevant captions for images. Combined, VQGAN-CLIP can take prompts from human input, and iterate to generate images that fit the prompts. Thanks to the generosity of creators sharing notebooks on Google Colab, the VQGAN-CLIP technique has seen widespread circulation. However, for regular usage across multiple sessions, I prefer a local setup that can be started up rapidly. Thus, this simple Streamlit app for generating VQGAN-CLIP images on a local environment. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 22
    gpt-2-simple

    gpt-2-simple

    Python package to easily retrain OpenAI's GPT-2 text-generating model

    A simple Python package that wraps existing model fine-tuning and generation scripts for OpenAI's GPT-2 text generation model (specifically the "small" 124M and "medium" 355M hyperparameter versions). Additionally, this package allows easier generation of text, generating to a file for easy curation, allowing for prefixes to force the text to start with a given phrase. For finetuning, it is strongly recommended to use a GPU, although you can generate using a CPU (albeit much more slowly). If...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 23
    Old Photo Restoration

    Old Photo Restoration

    Bringing Old Photo Back to Life (CVPR 2020 oral)

    We propose to restore old photos that suffer from severe degradation through a deep learning approach. Unlike conventional restoration tasks that can be solved through supervised learning, the degradation in real photos is complex and the domain gap between synthetic images and real old photos makes the network fail to generalize. Therefore, we propose a novel triplet domain translation network by leveraging real photos along with massive synthetic image pairs. Specifically, we train two...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    Opyrator

    Opyrator

    Turns your machine learning code into microservices with web API

    ...It cuts out all the pain for productizing and sharing your Python code - or anything you can wrap into a single Python function. An Opyrator-compatible function is required to have an input parameter and return value based on Pydantic models. The input and output models are specified via type hints. You can launch a graphical user interface - powered by Streamlit - for your compatible function. The UI is auto-generated from the input- and output-schema of the given function.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    CapsGNN

    CapsGNN

    A PyTorch implementation of "Capsule Graph Neural Network"

    A PyTorch implementation of "Capsule Graph Neural Network" (ICLR 2019). The high-quality node embeddings learned from the Graph Neural Networks (GNNs) have been applied to a wide range of node-based applications and some of them have achieved state-of-the-art (SOTA) performance. However, when applying node embeddings learned from GNNs to generate graph embeddings, the scalar node representation may not suffice to preserve the node/graph properties efficiently, resulting in sub-optimal graph...
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB