Showing 16 open source projects for "ps-scripts"

View related business solutions
  • Easily Host LLMs and Web Apps on Cloud Run Icon
    Easily Host LLMs and Web Apps on Cloud Run

    Run everything from popular models with on-demand NVIDIA L4 GPUs to web apps without infrastructure management.

    Run frontend and backend services, batch jobs, host LLMs, and queue processing workloads without the need to manage infrastructure. Cloud Run gives you on-demand GPU access for hosting LLMs and running real-time AI—with 5-second cold starts and automatic scale-to-zero so you only pay for actual usage. New customers get $300 in free credit to start.
    Try Cloud Run Free
  • $300 in Free Credit for Your Google Cloud Projects Icon
    $300 in Free Credit for Your Google Cloud Projects

    Build, test, and explore on Google Cloud with $300 in free credit. No hidden charges. No surprise bills.

    Launch your next project with $300 in free Google Cloud credit—no hidden charges. Test, build, and deploy without risk. Use your credit across the Google Cloud platform to find what works best for your needs. After your credits are used, continue building with free monthly usage products. Only pay when you're ready to scale. Sign up in minutes and start exploring.
    Start Free Trial
  • 1
    terminalGPT

    terminalGPT

    Get GPT like ChatGPT on your terminal

    ...Select + Create new secret key 4. Copy generated key Get started: Using tgpt: npm -g install terminalgpt or yarn global add terminalgpt Run tgpt chat ps.: If it is your first time running it, it will ask for open AI key , paste generated key from pre-requisite steps
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    YData Synthetic

    YData Synthetic

    Synthetic data generators for tabular and time-series data

    ...This repository contains material related to Generative Adversarial Networks for synthetic data generation, in particular regular tabular data and time-series. It consists a set of different GANs architectures developed using Tensorflow 2.0. Several example Jupyter Notebooks and Python scripts are included, to show how to use the different architectures. YData synthetic has now a UI interface to guide you through the steps and inputs to generate structure tabular data. The streamlit app is available form v1.0.0 onwards.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3

    uweb browser: unlimited power

    minimal suckless android web browser with unlimited power

    ... - Customizable: user-defined menus, (new) buttons and gestures for user agents, bookmarklets, url services, shell commands, internal functionality links and text processing etc. - Convenient: book/dictionary/txt/command line/app can be search engine. - Tiny: less than 200k - Fast: run fast, even with thousands of user provided css/scripts - Efficient: less touches, one click to reach any number of search engines without repeated input; automate online services. - URL bar command line support ("!" and .js files as commands). - user-defined site-specific JS/CSS/HTML/preprocessing. - Online play/preview/preprocess for downloadable resources. - Multiple type profiles: switch any data including logins/config orthogonally - web automation, crontab (alarm clock)
    Downloads: 10 This Week
    Last Update:
    See Project
  • 4
    gptee

    gptee

    LLMs done the UNIX-y way

    Output from a language model using standard input as the prompt. Now supporting GPT3.5 chat completions! gptee was designed for use within shell scripts and other programs and also works in interactive shells. You can compose commands and execute them in a script. Proceed with caution before running arbitrary shell scripts. Using a chat completion model (like gpt-3.5-turbo), you can then inject a system message with -s or --system messages. For davinci and other non-chat models, the output is prefixed to the prompt. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 99.99% Uptime for MySQL and PostgreSQL on Google Cloud Icon
    99.99% Uptime for MySQL and PostgreSQL on Google Cloud

    Enterprise Plus edition delivers sub-second maintenance downtime and 2x read/write performance. Built for critical apps.

    Cloud SQL Enterprise Plus gives you a 99.99% availability SLA with near-zero downtime maintenance—typically under 10 seconds. Get 2x better read/write performance, intelligent data caching, and 35 days of point-in-time recovery. Supports MySQL, PostgreSQL, and SQL Server with built-in vector search for gen AI apps. New customers get $300 in free credit.
    Try Cloud SQL Free
  • 5
    ChatGPT UI

    ChatGPT UI

    A ChatGPT web client that supports multiple users, and databases

    A ChatGPT web client that supports multiple users, multiple database connections for persistent data storage, supports i18n. Provides Docker images and quick deployment scripts. Support gpt-4 model. You can select the model in the "Model Parameters" of the front-end. The GPT-4 model requires whitelist access from OpenAI. Added web search capability to generate more relevant and up-to-date answers from ChatGPT! This feature is off by default, you can turn it on in `Chat->Settings` in the admin panel, there is a record `open_web_search` in Settings, set its value to True. ...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 6
    Edge GPT

    Edge GPT

    Reverse engineered API of Microsoft's Bing Chat

    EdgeGPT is a reverse-engineered client project that exposed a programmatic interface to Microsoft’s Bing Chat experience (often associated with “Edge” due to access patterns and requirements at the time). It was built to let developers run prompts and retrieve responses through scripts and applications, effectively turning the chat experience into something that could be automated and integrated into tools. The repository gained popularity because it provided a practical way to experiment with Bing Chat programmatically, including CLI-style usage patterns and developer-oriented documentation. As with many reverse-engineered clients, it depended on upstream behavior that could change, and the project’s lifecycle reflected that reality. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Point-E

    Point-E

    Point cloud diffusion for 3D model synthesis

    ...While it does not match the fine detail of some slower methods, the tradeoff in speed makes it practical for prototyping and interactive 3D generation. The repository includes inference scripts, utilities for converting point clouds to meshes (e.g. via signed distance function regression), sample notebooks, and weight checkpoints. It also provides documentation on limitations, usage instructions, and example outputs.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 8
    G-Diffuser Bot

    G-Diffuser Bot

    Discord bot and Interface for Stable Diffusion

    The first release of the all-in-one installer version of G-Diffuser is here. This release no longer requires the installation of WSL or Docker and has a systray icon to keep track of and launch G-Diffuser components. The infinite zoom scripts have been updated with some improvements, notably a new compositer script that is hundreds of times faster than before. The first release of the all-in-one installer is here. It notably features much easier "one-click" installation and updating, as well as a systray icon to keep track of g-diffuser programs and the server while it is running. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Minimal text diffusion

    Minimal text diffusion

    A minimal implementation of diffusion models for text generation

    ...The main idea was to retain just enough code to allow training a simple diffusion model and generating samples, remove image-related terms, and make it easier to use. To train a model, run scripts/train.sh. By default, this will train a model on the simple corpus. However, you can change this to any text file using the --train_data argument. Note that you may have to increase the sequence length (--seq_len) if your corpus is longer than the simple corpus. The other default arguments are set to match the best setting I found for the simple corpus.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Run Any Workload on Compute Engine VMs Icon
    Run Any Workload on Compute Engine VMs

    From dev environments to AI training, choose preset or custom VMs with 1–96 vCPUs and industry-leading 99.95% uptime SLA.

    Compute Engine delivers high-performance virtual machines for web apps, databases, containers, and AI workloads. Choose from general-purpose, compute-optimized, or GPU/TPU-accelerated machine types—or build custom VMs to match your exact specs. With live migration and automatic failover, your workloads stay online. New customers get $300 in free credits.
    Try Compute Engine
  • 10
    gpt-2-simple

    gpt-2-simple

    Python package to easily retrain OpenAI's GPT-2 text-generating model

    A simple Python package that wraps existing model fine-tuning and generation scripts for OpenAI's GPT-2 text generation model (specifically the "small" 124M and "medium" 355M hyperparameter versions). Additionally, this package allows easier generation of text, generating to a file for easy curation, allowing for prefixes to force the text to start with a given phrase. For finetuning, it is strongly recommended to use a GPU, although you can generate using a CPU (albeit much more slowly). ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Pipeline for training Language Models

    Pipeline for training Language Models

    Pipeline for training Language Models using PyTorch.

    Pipeline for training Language Models using PyTorch. Inspired by Yandex Data School NLP Course (week 03: Language Modeling) Prepared text file with space-separated words on each line.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    GPT2 for Multiple Languages

    GPT2 for Multiple Languages

    GPT2 for Multiple Languages, including pretrained models

    ...The contents in this repository are for academic research purpose, and we do not provide any conclusive remarks. Research supported with Cloud TPUs from Google's TensorFlow Research Cloud (TFRC) Simplifed GPT2 train scripts(based on Grover, supporting TPUs). Ported bert tokenizer, multilingual corpus compatible. 1.5B GPT2 pretrained Chinese model (~15G corpus, 10w steps). Batteries-included Colab demo. 1.5B GPT2 pretrained Chinese model (~30G corpus, 22w steps).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Image Super-Resolution (ISR)

    Image Super-Resolution (ISR)

    Super-scale your images and run experiments with Residual Dense

    The goal of this project is to upscale and improve the quality of low-resolution images. This project contains Keras implementations of different Residual Dense Networks for Single Image Super-Resolution (ISR) as well as scripts to train these networks using content and adversarial loss components. Docker scripts and Google Colab notebooks are available to carry training and prediction. Also, we provide scripts to facilitate training on the cloud with AWS and Nvidia-docker with only a few commands. When training your own model, start with only PSNR loss (50+ epochs, depending on the dataset) and only then introduce GANS and feature loss. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 14
    GPT-2 FR

    GPT-2 FR

    GPT-2 French demo | Démo française de GPT-2

    OpenAI GPT-2 model trained on four different datasets in French. Books in French, French film scripts, reports of parliamentary debates, Tweet by Emmanuel Macron, allowing to generate text. Tensorflow and gpt-2-simple are required in order to fine-tune GPT-2. Create an environment then install the two packages pip install tensorflow==1.14 gpt-2-simple. A script and a notebook are available in the src folder to fine-tune GPT-2 on your own datasets.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    PyTorch pretrained BigGAN

    PyTorch pretrained BigGAN

    PyTorch implementation of BigGAN with pretrained weights

    ...This repository contains an op-for-op PyTorch reimplementation of DeepMind's BigGAN that was released with the paper Large Scale GAN Training for High Fidelity Natural Image Synthesis. This PyTorch implementation of BigGAN is provided with the pretrained 128x128, 256x256 and 512x512 models by DeepMind. We also provide the scripts used to download and convert these models from the TensorFlow Hub models. This reimplementation was done from the raw computation graph of the Tensorflow version and behave similarly to the TensorFlow version (variance of the output difference of the order of 1e-5). This implementation currently only contains the generator as the weights of the discriminator were not released (although the structure of the discriminator is very similar to the generator so it could be added pretty easily.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 16
    Finetune Transformer LM

    Finetune Transformer LM

    Code for "Improving Language Understanding by Generative Pre-Training"

    ...It documents that runs are non-deterministic due to certain GPU operations and reports a median accuracy over multiple trials that is slightly below the single-run result in the paper, reflecting expected variance in practice. The project ships lightweight training, data, and analysis scripts, keeping the footprint small while making the experimental pipeline transparent. It is provided as archived, research-grade code intended for replication and study rather than continuous development.
    Downloads: 5 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB