520 projects for "parrot-security-4" with 2 filters applied:

  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • AI-generated apps that pass security review Icon
    AI-generated apps that pass security review

    Stop waiting on engineering. Build production-ready internal tools with AI—on your company data, in your cloud.

    Retool lets you generate dashboards, admin panels, and workflows directly on your data. Type something like “Build me a revenue dashboard on my Stripe data” and get a working app with security, permissions, and compliance built in from day one. Whether on our cloud or self-hosted, create the internal software your team needs without compromising enterprise standards or control.
    Try Retool free
  • 1
    Agently 4

    Agently 4

    Build GenAI application quick and easy

    Agently is a Python framework for building generative-AI (“GenAI”) applications; it focuses on enabling developers to orchestrate AI agents, workflows, and event-driven logic in a robust, reusable way. With Agently, one can define agents that call different models, chain tasks, trigger workflows based on events, and switch models with minimal code changes. It abstracts away boilerplate around model API calls, tool usage, prompt management, and workflow state. The project aims at...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    GLM-4.6

    GLM-4.6

    Agentic, Reasoning, and Coding (ARC) foundation models

    ...GLM-4.6 also enhances writing quality, producing outputs that better align with human preferences and role-playing scenarios. Benchmark evaluations demonstrate that it not only outperforms GLM-4.5 but also rivals leading global models such as DeepSeek-V3.1-Terminus and Claude Sonnet 4.
    Downloads: 79 This Week
    Last Update:
    See Project
  • 3
    GLM-4.7

    GLM-4.7

    Advanced language and coding AI model

    GLM-4.7 is an advanced agent-oriented large language model designed as a high-performance coding and reasoning partner. It delivers significant gains over GLM-4.6 in multilingual agentic coding, terminal-based workflows, and real-world developer benchmarks such as SWE-bench and Terminal Bench 2.0. The model introduces stronger “thinking before acting” behavior, improving stability and accuracy in complex agent frameworks like Claude Code, Cline, and Roo Code. GLM-4.7 also advances “vibe...
    Downloads: 75 This Week
    Last Update:
    See Project
  • 4
    GLM-4.6V

    GLM-4.6V

    GLM-4.6V/4.5V/4.1V-Thinking, towards versatile multimodal reasoning

    GLM-4.6V represents the latest generation of the GLM-V family and marks a major step forward in multimodal AI by combining advanced vision-language understanding with native “tool-call” capabilities, long-context reasoning, and strong generalization across domains. Unlike many vision-language models that treat images and text separately or require intermediate conversions, GLM-4.6V allows inputs such as images, screenshots or document pages directly as part of its reasoning pipeline — and...
    Downloads: 2 This Week
    Last Update:
    See Project
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 5
    GLM-4.5

    GLM-4.5

    GLM-4.5: Open-source LLM for intelligent agents by Z.ai

    GLM-4.5 is a cutting-edge open-source large language model designed by Z.ai for intelligent agent applications. The flagship GLM-4.5 model has 355 billion total parameters with 32 billion active parameters, while the compact GLM-4.5-Air version offers 106 billion total parameters and 12 billion active parameters. Both models unify reasoning, coding, and intelligent agent capabilities, providing two modes: a thinking mode for complex reasoning and tool usage, and a non-thinking mode for...
    Downloads: 50 This Week
    Last Update:
    See Project
  • 6
    GLM-4.5V

    GLM-4.5V

    GLM-4.6V/4.5V/4.1V-Thinking, towards versatile multimodal reasoning

    GLM-4.5V is the preceding iteration in the GLM-V series that laid much of the groundwork for general multimodal reasoning and vision-language understanding. It embodies the design philosophy of mixing visual and textual modalities into a unified model capable of general-purpose reasoning, content understanding, and generation, while already supporting a wide variety of tasks: from image captioning and visual question answering to content recognition, GUI-based agents, video understanding,...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 7
    GLM-4.1V

    GLM-4.1V

    GLM-4.6V/4.5V/4.1V-Thinking, towards versatile multimodal reasoning

    GLM-4.1V — often referred to as a smaller / lighter version of the GLM-V family — offers a more resource-efficient option for users who want multimodal capabilities without requiring large compute resources. Though smaller in scale, GLM-4.1V maintains competitive performance, particularly impressive on many benchmarks for models of its size: in fact, on a number of multimodal reasoning and vision-language tasks it outperforms some much larger models from other families. It represents a...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Claude Code Security Reviewer

    Claude Code Security Reviewer

    An AI-powered security review GitHub Action using Claude

    The claude-code-security-review repository implements a GitHub Action that uses Claude (via the Anthropic API) to perform semantic security audits of code changes in pull requests. Rather than relying purely on pattern matching or static analysis, this action feeds diffs and surrounding context to Claude to reason about potential vulnerabilities (e.g. injection, misconfigurations, secrets exposure, etc).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    PentestGPT

    PentestGPT

    Automated Penetration Testing Agentic Framework Powered by LLMs

    PentestGPT is an AI-powered autonomous penetration testing agent designed to perform intelligent, end-to-end security assessments using large language models. Published at USENIX Security 2024, it combines advanced reasoning with an agentic workflow to automate tasks traditionally handled by human pentesters. The platform supports multiple penetration testing categories, including web security, cryptography, reversing, forensics, privilege escalation, and binary exploitation. ...
    Downloads: 307 This Week
    Last Update:
    See Project
  • MyQ Print Management Software Icon
    MyQ Print Management Software

    SAVE TIME WITH PERSONALIZED PRINT SOLUTIONS

    Boost your digital or traditional workplace with MyQ’s secure print and scan solutions that respect your time and help you focus on what you do best.
    Learn More
  • 10
    Strix

    Strix

    Open-source AI hackers to find and fix your app’s vulnerabilities

    ...The platform is intended for developers and security teams that need rapid security assessments without the overhead of manual penetration testing engagements. Strix can orchestrate multiple cooperating agents that divide investigation tasks and collaboratively analyze complex applications or infrastructure.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 11
    CyberStrikeAI

    CyberStrikeAI

    CyberStrikeAI is an AI-native security testing platform built in Go

    CyberStrikeAI is an AI-native security testing platform built in Go that brings autonomous penetration testing, vulnerability discovery, and attack chain analysis into a unified interface. The platform integrates over 100 security tools out of the box and pairs them with an intelligent orchestration engine that can be directed via natural language or policy definitions, allowing users to automate reconnaissance, scanning, exploitation, and reporting without manual sequencing of tools. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 12
    Sec-Context

    Sec-Context

    AI Code Security Anti-Patterns distilled from 150+ sources

    Sec-Context is a curated security research project that distills common code anti-patterns and vulnerabilities that generative AI tends to produce, presenting them as a comprehensive set of examples and secure alternatives that can be used to train or guide AI assistants and reviewers toward safer code generation. It compiles insights from over 150 industry and academic sources into structured reference documents that outline real-world security problems such as hardcoded secrets, SQL injection, cross-site scripting, command injection, weak password storage, and other frequent issues that occur when code is auto-generated without context of best practices. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 13
    Open WebUI

    Open WebUI

    User-friendly AI Interface

    ...Key features include effortless setup via Docker or Kubernetes, seamless integration with OpenAI-compatible APIs, granular permissions and user groups for enhanced security, responsive design across devices, and full Markdown and LaTeX support for enriched interactions. Additionally, Open WebUI offers a Progressive Web App (PWA) for mobile devices, providing offline access and a native app-like experience. The platform also includes a Model Builder, allowing users to create custom models from base Ollama models directly within the interface. ...
    Downloads: 88 This Week
    Last Update:
    See Project
  • 14
    Vulnhuntr

    Vulnhuntr

    AI tool for detecting complex vulnerabilities in Python codebases

    ...Vulnhuntr can generate detailed findings, including vulnerability explanations and potential exploit paths, helping developers and security teams understand risks faster. It supports multiple LLM providers such as OpenAI, Anthropic, and Ollama, and can be run via CLI, Docker, or pipx. Vulnhuntr is particularly useful for early-stage security reviews, bug bounty hunting, and auditing dependencies for hidden risks across open source projects.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 15
    Skill Scanner

    Skill Scanner

    Security Scanner for Agent Skills

    This repository is a public security-focused scanning tool intended to analyze and assess AI agent skills for potential issues, quality concerns, and vulnerabilities. It acts as a scanner that inspects Agent Skills packages to flag structural problems, inconsistencies, or security flaws before they are deployed or integrated into agent workflows. Because agent skills can contain executable instructions and logic, scanning them for risky patterns is essential to prevent inadvertent exploitation when used by intelligent systems. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 16
    Monkey Code

    Monkey Code

    Enterprise-grade AI programming assistant designed for R&D collab

    ...One of its defining characteristics is its support for private deployment and fully offline operation, which makes it especially suitable for organizations with strict data privacy or security requirements. The system includes a comprehensive management panel that allows teams to audit, monitor, and control how AI participates in coding workflows, ensuring accountability and governance at scale. MonkeyCode also integrates automated code security scanning to detect vulnerabilities in both human-written and AI-generated code, reinforcing secure development practices.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 17
    AICGSecEval

    AICGSecEval

    A.S.E (AICGSecEval) is a repository-level AI-generated code security

    AICGSecEval is an open-source benchmark framework designed to evaluate the security of code generated by artificial intelligence systems. The project was developed to address concerns that AI-assisted programming tools may produce insecure code containing vulnerabilities such as injection flaws or unsafe logic. The framework constructs evaluation tasks based on real-world software repositories and known vulnerability cases derived from CVE records.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    LLM Guard

    LLM Guard

    The Security Toolkit for LLM Interactions

    LLM Guard is an open-source security toolkit designed to protect large language model applications from various security risks and adversarial attacks. The library acts as a protective layer between users and language models by analyzing inputs and outputs before they reach or leave the model. It includes scanning mechanisms that detect malicious prompts, prompt injection attempts, toxic content, and other harmful inputs that could compromise AI systems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    Trail of Bits Skills Marketplace

    Trail of Bits Skills Marketplace

    Trail of Bits Claude Code skills for security research, vulnerability

    ...This project leverages the agent skills architecture to let AI assistants take on detailed, repeatable security procedures that are typically manual, such as parsing Burp Suite projects or conducting variant analysis across codebases.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    DeepAudit

    DeepAudit

    AI multi-agent platform for automated code security auditing system

    DeepAudit is an open source code security auditing platform that uses a multi-agent architecture to analyze and identify vulnerabilities in software projects. Instead of relying solely on traditional static analysis, it simulates the reasoning process of security experts through coordinated agents responsible for orchestration, reconnaissance, analysis, and verification.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 21
    Open Semantic Search

    Open Semantic Search

    Open source semantic search and text analytics for large document sets

    Open Semantic Search is an open source research and analytics platform designed for searching, analyzing, and exploring large collections of documents using semantic search technologies. It provides an integrated search server combined with a document processing pipeline that supports crawling, text extraction, and automated analysis of content from many different sources. Open Semantic Search includes an ETL framework that can ingest documents, process them through analysis steps, and...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 22
    Open Gauss

    Open Gauss

    Project-scoped Lean workflow orchestrator from Math, Inc.

    Open Gauss is an enterprise-grade open-source relational database management system designed to handle large-scale data processing with high performance, reliability, and security. It is based on the PostgreSQL ecosystem but significantly extends its capabilities through architectural optimizations, AI-driven features, and enterprise-level enhancements. The database organizes data using the relational model, storing structured information in tables composed of rows and columns while supporting standard SQL for querying and management. ...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 23
    FuzzyAI Fuzzer

    FuzzyAI Fuzzer

    A powerful tool for automated LLM fuzzing

    FuzzyAI is an open-source fuzzing framework designed to test the security and reliability of large language model applications. The tool automates the process of generating adversarial prompts and input variations to identify vulnerabilities such as jailbreaks, prompt injections, or unsafe model responses. It allows developers and security researchers to systematically evaluate the robustness of LLM-based systems by simulating a wide range of malicious or unexpected inputs. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    BruteForceAI

    BruteForceAI

    Advanced LLM-powered brute-force tool combining AI intelligence

    BruteForceAI is an open-source security testing tool that applies large language models to the analysis of login forms and authentication flows in web applications. At a high level, the project uses AI to inspect HTML content, identify the relevant form elements, and automate selector discovery so that a tester does not need to hand-map every field before evaluation.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 25
    Claude Code Skills & Plugins Hub

    Claude Code Skills & Plugins Hub

    270+ Claude Code plugins with 739 agent skills

    ...The repository functions as a marketplace-style collection of hundreds of plugins and specialized skills that enable Claude Code to perform complex development, automation, and operational tasks. These plugins cover a wide range of domains including DevOps automation, security testing, API debugging, infrastructure management, and AI workflow orchestration. The project also includes orchestration patterns and best practices that guide how multiple AI agents or tools can collaborate effectively in software development workflows. Developers can install plugins through a package-style plugin system and integrate them with their Claude Code environment using standardized commands.
    Downloads: 3 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • 5
  • Next
MongoDB Logo MongoDB