WebLLM
WebLLM is a high-performance, in-browser language model inference engine that leverages WebGPU for hardware acceleration, enabling powerful LLM operations directly within web browsers without server-side processing. It offers full OpenAI API compatibility, allowing seamless integration with functionalities such as JSON mode, function-calling, and streaming. WebLLM natively supports a range of models, including Llama, Phi, Gemma, RedPajama, Mistral, and Qwen, making it versatile for various AI tasks. Users can easily integrate and deploy custom models in MLC format, adapting WebLLM to specific needs and scenarios. The platform facilitates plug-and-play integration through package managers like NPM and Yarn, or directly via CDN, complemented by comprehensive examples and a modular design for connecting with UI components. It supports streaming chat completions for real-time output generation, enhancing interactive applications like chatbots and virtual assistants.
Learn more
MindMac
MindMac is a native macOS application designed to enhance productivity by integrating seamlessly with ChatGPT and other AI models. It supports multiple AI providers, including OpenAI, Azure OpenAI, Google AI with Gemini, Gemini Enterprise Agent Platform, Anthropic Claude, OpenRouter, Mistral AI, Cohere, Perplexity, OctoAI, and local LLMs via LMStudio, LocalAI, GPT4All, Ollama, and llama.cpp. MindMac offers over 150 built-in prompt templates to facilitate user interaction and allows for extensive customization of OpenAI parameters, appearance, context modes, and keyboard shortcuts. The application features a powerful inline mode, enabling users to generate content or ask questions within any application without switching windows. MindMac ensures privacy by storing API keys securely in the Mac's Keychain and sending data directly to the AI provider without intermediary servers. The app is free to use with basic features, requiring no account for setup.
Learn more
Note67
Note67 is a privacy-centric meeting assistant designed for professionals who demand total control over their data. Unlike traditional transcription tools that rely on cloud processing, Note67 is an open-source, local-first application for macOS that captures audio, transcribes speech, and generates intelligent summaries entirely on your device. No audio or text ever leaves your machine, ensuring zero data leakage. Built with performance and security in mind, the application leverages the power of Rust and Tauri to deliver a lightweight, native experience. It integrates seamless local AI capabilities, utilizing Whisper for high-accuracy speech-to-text and Ollama for generating insightful meeting summaries using local Large Language Models (LLMs). Key Features: 100% Local Processing: Powered by on-device Whisper models, ensuring your audio and transcripts remain completely private.
Learn more
Locally AI
Locally AI is an on-device AI application that allows users to run powerful language models directly on their iPhone, iPad, or Mac without relying on cloud infrastructure or an internet connection. Built on Apple’s MLX framework, it delivers fast, efficient performance while minimizing power usage, enabling a seamless experience for chatting, creating, learning, and exploring AI capabilities across devices. It supports multiple open models such as Llama, Gemma, Qwen, and DeepSeek, allowing users to switch between them and tailor outputs to different tasks. Everything runs entirely offline, meaning no login is required, and no data is collected or transmitted, ensuring complete privacy and control over personal information. Users can interact with AI through natural conversations, analyze documents or images, and generate text in a unified interface designed for simplicity and responsiveness.
Learn more