49 projects for "artificial intelligence python" with 2 filters applied:

  • Cloud-based help desk software with ServoDesk Icon
    Cloud-based help desk software with ServoDesk

    Full access to Enterprise features. No credit card required.

    What if You Could Automate 90% of Your Repetitive Tasks in Under 30 Days? At ServoDesk, we help businesses like yours automate operations with AI, allowing you to cut service times in half and increase productivity by 25% - without hiring more staff.
    Try ServoDesk for free
  • Customize SharePoint Without Coding Icon
    Customize SharePoint Without Coding

    QuickApps is an innovative No-Code solution that can completely transform your SharePoint experience and help maximize your ROI.

    QuickApps is a set of powerful SharePoint web apps that lets you build applications that are easily supported, maintained and upgraded, ensuring their long-term impact and return on investment.
    Learn More
  • 1
    BIG-bench

    BIG-bench

    Beyond the Imitation Game collaborative benchmark for measuring

    BIG-bench (Beyond the Imitation Game Benchmark) is a large, collaborative benchmark suite designed to probe the capabilities and limitations of large language models across hundreds of diverse tasks. Rather than focusing on a single metric or domain, it aggregates many hand-authored tasks that test reasoning, commonsense, math, linguistics, ethics, and creativity. Tasks are intentionally heterogeneous: some are multiple-choice with exact scoring, others are free-form generation judged by...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    Automated Interpretability

    Automated Interpretability

    Code for Language models can explain neurons in language models paper

    The automated-interpretability repository implements tools and pipelines for automatically generating, simulating, and scoring explanations of neuron (or latent feature) behavior in neural networks. Instead of relying purely on manual, ad hoc interpretability probing, this repo aims to scale interpretability by using algorithmic methods that produce candidate explanations and assess their quality. It includes a “neuron explainer” component that, given a target neuron or latent feature,...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Evals

    Evals

    Evals is a framework for evaluating LLMs and LLM systems

    The openai/evals repository is a framework and registry for evaluating large language models and systems built with LLMs. It’s designed to let you define “evals” (evaluation tasks) in a structured way and run them against different models or agents, with the ability to score, compare, and analyze results. The framework supports templated YAML eval definitions, solver-based evaluations, custom metrics, and composition of multi-step evaluations. It includes utilities and APIs to plug in...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    CSGHub

    CSGHub

    CSGHub is a brand-new open-source platform for managing LLMs

    CSGHub is an open-source framework designed for collaborative scientific research and content generation. It enables researchers to utilize AI-driven tools for literature review, hypothesis generation, and automated writing assistance, streamlining the scientific discovery process.
    Downloads: 2 This Week
    Last Update:
    See Project
  • ManageEngine ADAudit Plus | A UBA-driven change auditor Icon
    ManageEngine ADAudit Plus | A UBA-driven change auditor

    ADAudit Plus helps keep your Active Directory, file servers, Windows servers and workstations secure and compliant.

    ADAudit Plus helps keep your Windows Server ecosystem secure and compliant by providing full visibility into all activities. ADAudit Plus provides a clear picture of all changes made to your AD resources including AD objects and their attributes, group policy, and more.
    Learn More
  • 5
    Dramatron

    Dramatron

    Dramatron uses large language models to generate coherent scripts

    Dramatron is an interactive co-writing tool developed by Google DeepMind that leverages large language models to help authors create screenplays and theatre scripts. It uses a hierarchical story generation approach to maintain coherence and structure across multiple levels of a narrative, from a single logline to detailed character descriptions, locations, plot points, and dialogue. Dramatron operates as a creative assistant rather than a fully autonomous system, offering human writers...
    Downloads: 7 This Week
    Last Update:
    See Project
  • 6
    Granite 3.0 Language Models

    Granite 3.0 Language Models

    New set of lightweight state-of-the-art, open foundation models

    This repository introduces Granite 3.0 language models as lightweight, state-of-the-art open foundation models built to natively support multilinguality, coding, reasoning, and tool usage. A central goal is efficient deployment, including the potential to run on constrained compute resources while remaining useful for a broad span of enterprise tasks. The repo positions the models for both research and commercial use under an Apache-2.0 license, signaling permissive adoption paths....
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Prompt Engineering Techniques

    Prompt Engineering Techniques

    Collection of tutorials for Prompt Engineering techniques

    Prompt Engineering Techniques is a focused companion repository that teaches prompt engineering systematically, from fundamentals to advanced strategies. It contains around twenty-plus hands-on Jupyter notebooks, each dedicated to a specific technique such as basic prompt structures, prompt templates and variables, zero-shot prompting, few-shot prompting, chain-of-thought, self-consistency, constrained generation, role prompting, task decomposition, and more. The tutorials are designed to be...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 8
    LLM Datasets

    LLM Datasets

    Curated list of datasets and tools for post-training

    LLM Datasets curates and standardizes datasets commonly used to train and fine-tune large language models, reducing the overhead of hunting down sources and normalizing formats. The repository aims to make datasets easy to inspect and transform, with scripts for downloading, deduping, cleaning, and converting to formats like JSONL that slot into training pipelines. It highlights instruction-tuning and conversation-style corpora while also pointing to code, math, or domain-specific sets for...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 9
    DeepSeek LLM

    DeepSeek LLM

    DeepSeek LLM: Let there be answers

    The DeepSeek-LLM repository hosts the code, model files, evaluations, and documentation for DeepSeek’s LLM series (notably the 67B Chat variant). Its tagline is “Let there be answers.” The repo includes an “evaluation” folder (with results like math benchmark scores) and code artifacts (e.g. pre-commit config) that support model development and deployment. According to the evaluation files, DeepSeek LLM 67B Chat achieves strong performance on math benchmarks under both chain-of-thought (CoT)...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Case Management Software for Nonprofits Icon
    Case Management Software for Nonprofits

    CharityTracker offers easy-to-use case management software for nonprofits, helping them track assistance, measure outcomes, and generate reports.

    A case management CRM that focuses on your clients, to measure real-world impact, not just data. Your Day Just Got Shorter - We know you’re busy and don’t have time for complicated software – We’ve changed that! Even your staff will enjoy using the system – No more headaches and missing data from reports so you can use your valuable time to serve more people. CharityTracker is a cloud-based, HIPAA compliant software that’s easy to use for nonprofits of all sizes. With no setup fees, hidden fees, or contracts, you can cancel anytime! Our time saving solution provides a secure way to streamline how you manage cases, record case notes & assistance, create reports, collaborate, set goals, measure outcomes, make referrals & prove your impact with change-over-time reports for your funders. We're at work in 2,200 communities! Book a free demo. Pricing starts at just $20/mo. user
    Learn More
  • 10
    GenAI Agents

    GenAI Agents

    Implementations for various Generative AI Agent techniques

    GenAI Agents is a large, tutorial-driven repository that teaches you how to design, build, and experiment with generative AI agents. It spans a spectrum from simple conversational bots and basic question-answering agents to complex multi-agent systems that coordinate on research, education, business workflows, and creative tasks. The implementations leverage modern frameworks such as LangChain, LangGraph, AutoGen, PydanticAI, CrewAI, and more, showing how each can be wired into realistic...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 11
    Secret Llama

    Secret Llama

    Fully private LLM chatbot that runs entirely with a browser

    Secret Llama is a privacy-first large-language-model chatbot that runs entirely inside your web browser, meaning no server is required and your conversation data never leaves your device. It focuses on open-source model support, letting you load families like Llama and Mistral directly in the client for fully local inference. Because everything happens in-browser, it can work offline once models are cached, which is helpful for air-gapped environments or travel. The interface mirrors the...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 12
    LLM Course

    LLM Course

    Course to get into Large Language Models (LLMs)

    LLM Course is a hands-on, notebook-driven path for learning how large language models work in practice, from data curation to training, fine-tuning, evaluating, and deploying. It emphasizes reproducible experiments: each step is demonstrated with runnable code, clear dependencies, and references to commonly used open-source models and libraries. Learners get exposure to multiple adaptation strategies—LoRA/QLoRA, instruction fine-tuning, and alignment techniques—so they can choose approaches...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 13
    Grok-1

    Grok-1

    Open-source, high-performance Mixture-of-Experts large language model

    Grok-1 is a 314-billion-parameter Mixture-of-Experts (MoE) large language model developed by xAI. Designed to optimize computational efficiency, it activates only 25% of its weights for each input token. In March 2024, xAI released Grok-1's model weights and architecture under the Apache 2.0 license, making them openly accessible to developers. The accompanying GitHub repository provides JAX example code for loading and running the model. Due to its substantial size, utilizing Grok-1...
    Downloads: 12 This Week
    Last Update:
    See Project
  • 14
    FinGPT

    FinGPT

    Open-Source Financial Large Language Models!

    FinGPT is an open-source large language model tailored specifically for financial tasks. Developed by AI4Finance Foundation, it is designed to assist with various financial applications, such as forecasting, financial sentiment analysis, and portfolio management. FinGPT has been trained on a diverse range of financial datasets, making it a powerful tool for finance professionals looking to leverage AI for data-driven decision-making. The model is freely available on platforms like Hugging...
    Leader badge
    Downloads: 13 This Week
    Last Update:
    See Project
  • 15
    Qwen2.5-Coder

    Qwen2.5-Coder

    Qwen2.5-Coder is the code version of Qwen2.5, the large language model

    Qwen2.5-Coder, developed by QwenLM, is an advanced open-source code generation model designed for developers seeking powerful and diverse coding capabilities. It includes multiple model sizes—ranging from 0.5B to 32B parameters—providing solutions for a wide array of coding needs. The model supports over 92 programming languages and offers exceptional performance in generating code, debugging, and mathematical problem-solving. Qwen2.5-Coder, with its long context length of 128K tokens, is...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 16
    ChatGLM2-6B

    ChatGLM2-6B

    An Open Bilingual Chat LLM | Open Source Bilingual Conversation LLM

    ChatGLM2-6B is an advanced open-source bilingual dialogue model developed by THUDM. It is the second iteration of the ChatGLM series, designed to offer enhanced performance while maintaining the strengths of its predecessor, including smooth conversation flow and low deployment barriers. The model is fine-tuned for both Chinese and English languages, making it a versatile tool for various multilingual applications. ChatGLM2-6B aims to push the boundaries of natural language understanding and...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Qwen2.5

    Qwen2.5

    Open source large language model by Alibaba

    Qwen2.5 is a series of large language models developed by the Qwen team at Alibaba Cloud, designed to enhance natural language understanding and generation across multiple languages. The models are available in various sizes, including 0.5B, 1.5B, 3B, 7B, 14B, 32B, and 72B parameters, catering to diverse computational requirements. Trained on a comprehensive dataset of up to 18 trillion tokens, Qwen2.5 models exhibit significant improvements in instruction following, long-text generation...
    Downloads: 27 This Week
    Last Update:
    See Project
  • 18
    GLM-130B

    GLM-130B

    GLM-130B: An Open Bilingual Pre-Trained Model (ICLR 2023)

    GLM-130B is an open bilingual (English and Chinese) dense language model with 130 billion parameters, released by the Tsinghua KEG Lab and collaborators as part of the General Language Model (GLM) series. It is designed for large-scale inference and supports both left-to-right generation and blank filling, making it versatile across NLP tasks. Trained on over 400 billion tokens (200B English, 200B Chinese), it achieves performance surpassing GPT-3 175B, OPT-175B, and BLOOM-176B on multiple...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 19
    FastEdit

    FastEdit

    Editing large language models within 10 seconds

    FastEdit focuses on rapid “model editing,” letting you surgically update facts or behaviors in an LLM without full fine-tuning. It implements practical editing algorithms that insert or revise knowledge with targeted parameter updates, aiming to preserve model quality outside the edited scope. This approach is valuable when you need urgent corrections—think product names, APIs, or fast-changing facts—without retraining on large corpora. The repository provides evaluation harnesses so you can...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 20
    GPT-NeoX

    GPT-NeoX

    Implementation of model parallel autoregressive transformers on GPUs

    This repository records EleutherAI's library for training large-scale language models on GPUs. Our current framework is based on NVIDIA's Megatron Language Model and has been augmented with techniques from DeepSpeed as well as some novel optimizations. We aim to make this repo a centralized and accessible place to gather techniques for training large-scale autoregressive language models, and accelerate research into large-scale training. For those looking for a TPU-centric codebase, we...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 21
    LM Human Preferences

    LM Human Preferences

    Code for the paper Fine-Tuning Language Models from Human Preferences

    lm-human-preferences is the official OpenAI codebase that implements the method from the paper Fine-Tuning Language Models from Human Preferences. Its purpose is to show how to align language models with human judgments by training a reward model from human comparisons and then fine-tuning a policy model using that reward signal. The repository includes scripts to train the reward model (learning to rank or score pairs of outputs), and to fine-tune a policy (a language model) with...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    Alpa

    Alpa

    Training and serving large-scale neural networks

    Alpa is a system for training and serving large-scale neural networks. Scaling neural networks to hundreds of billions of parameters has enabled dramatic breakthroughs such as GPT-3, but training and serving these large-scale neural networks require complicated distributed system techniques. Alpa aims to automate large-scale distributed training and serving with just a few lines of code.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Following Instructions with Feedback

    Following Instructions with Feedback

    Training Language Models to Follow Instructions with Human Feedback

    The following-instructions-human-feedback repository contains the code and supplementary materials underpinning OpenAI’s work in training language models (InstructGPT models) that better follow user instructions through human feedback. The repo hosts the model card, sample automatic evaluation outputs, and labeling guidelines used in the process. It is explicitly tied to the “Training language models to follow instructions with human feedback” paper, and serves as a reference for how OpenAI...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    Grade School Math

    Grade School Math

    8.5K high quality grade school math problems

    The grade-school-math repository (sometimes called GSM8K) is a curated dataset of 8,500 high-quality grade school math word problems intended for evaluating mathematical reasoning capabilities of language models. It is structured into 7,500 training problems and 1,000 test problems. These aren’t trivial exercises — many require multi-step reasoning, combining arithmetic operations, and handling intermediate steps (e.g. “If she sold half as many in May… how many in total?”). The problems are...
    Downloads: 0 This Week
    Last Update:
    See Project