Showing 2 open source projects for "chat script"

View related business solutions
  • Forever Free Full-Stack Observability | Grafana Cloud Icon
    Forever Free Full-Stack Observability | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 1
    Feishu ChatGPT

    Feishu ChatGPT

    Voice dialogue, role-playing, multi-topic discussion, picture creation

    Feishu × (GPT-3.5 + DALL·E + Whisper) = flying-like work experience. Voice dialogue, role-playing, multi-topic discussion, picture creation, table analysis, document export. Golang language, it goes without saying! Master the gin framework proficiently, developing the backend is as natural as breathing! Familiar with the SDKs of DingTalk, Feishu, Qiwei and other platforms, and be able to develop and integrate a series of amazing functions! Proficient in platform-based detail thinking, let...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 2
    Alpaca.cpp

    Alpaca.cpp

    Locally run an Instruction-Tuned Chat-Style LLM

    ...This combines the LLaMA foundation model with an open reproduction of Stanford Alpaca a fine-tuning of the base model to obey instructions (akin to the RLHF used to train ChatGPT) and a set of modifications to llama.cpp to add a chat interface. Download the zip file corresponding to your operating system from the latest release. The weights are based on the published fine-tunes from alpaca-lora, converted back into a PyTorch checkpoint with a modified script and then quantized with llama.cpp the regular way.
    Downloads: 2 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB