Showing 38 open source projects for "depth-violet"

View related business solutions
  • Enterprise-grade ITSM, for every business Icon
    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity.

    Freshservice is an intuitive, AI-powered platform that helps IT, operations, and business teams deliver exceptional service without the usual complexity. Automate repetitive tasks, resolve issues faster, and provide seamless support across the organization. From managing incidents and assets to driving smarter decisions, Freshservice makes it easy to stay efficient and scale with confidence.
    Try it Free
  • Go From AI Idea to AI App Fast Icon
    Go From AI Idea to AI App Fast

    One platform to build, fine-tune, and deploy ML models. No MLOps team required.

    Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
    Try Free
  • 1
    katana

    katana

    Fast CLI web crawler for discovering endpoints in modern web apps

    ...Through headless browsing, it can analyze dynamic content and single-page applications built with modern frameworks, improving its ability to uncover hidden paths and assets. Katana offers flexible configuration options such as depth control, concurrency limits, and filtering mechanisms to refine results and manage scanning scope.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 2
    crwlr

    crwlr

    Library for Rapid (Web) Crawler and Scraper Development

    ...A crawler could just load actually all links it is finding (and is allowed to load according to the robots.txt file), then it would just load the whole internet (if the URL(s) it starts with are no dead end). Or it can be restricted to load only links matching certain criteria (on same domain/host, URL path starts with "/foo",...) or only to a certain depth. A depth of 3 means 3 levels deep. Links found on the initial URLs provided to the crawler are level 1 and so on.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 3
    watercrawl

    watercrawl

    AI-ready web crawler that extracts and structures website content

    ...It enables developers and researchers to crawl web pages, extract meaningful information, and convert it into formats that are easier to process and analyze. It provides a modern crawling system that can automatically navigate links, control crawl depth, and collect content from targeted sections of a website. WaterCrawl supports customizable extraction rules so users can focus only on relevant elements while ignoring unnecessary page components. WaterCrawl also offers real-time monitoring capabilities, allowing users to track crawling progress, performance metrics, and errors during large data collection jobs. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Scrapy

    Scrapy

    A fast, high-level web crawling and web scraping framework

    Scrapy is a fast, open source, high-level framework for crawling websites and extracting structured data from these websites. Portable and written in Python, it can run on Windows, Linux, macOS and BSD. Scrapy is powerful, fast and simple, and also easily extensible. Simply write the rules to extract the data, and add new functionality if you wish without having to touch the core. Scrapy does the rest, and can be used in a number of applications. It can be used for data mining, monitoring...
    Downloads: 20 This Week
    Last Update:
    See Project
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 5
    SimpRead

    SimpRead

    An extension that allows you to instantly enter immersive reading

    Unique function, by extracting titles, descriptions, text, media pictures/videos and other resources to generate pages that conform to Chinese reading habits, with zero-interference and immersive features, suitable for in-depth reading. Do not change the structure of the current page, just highlight the part that needs to be read, suitable for temporary reading or websites that are not adapted to the reading mode. Have pictures you don’t want to show? no problem! Any element on the page can be re-edited. The brand-new lexical analysis engine can not only recognize Wordpress, Hexo, Ghost, and Discuz, as long as it is a well-structured page, it can automatically generate a reading mode. ...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 6
    crawley

    crawley

    The unix-way web crawler

    ...Grabs most of useful resources URLs (pics, videos, audios, forms, etc...) Found URLs are streamed to stdout and guaranteed to be unique (with fragments omitted) Scan depth (limited by starting host and path, by default - 0) can be configured. Can crawl rules and sitemaps from robots.txt. Brute mode - scan HTML comments for URLs (this can lead to bogus results) Make use of HTTP_PROXY / HTTPS_PROXY environment values + handle proxy auth. Directory-only scan mode (aka fast-scan)
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Spider

    Spider

    High-performance Rust web crawler and scraper for large-scale data

    ...Spider also provides mechanisms for subscribing to crawl events so developers can process page data such as URLs, status codes, or HTML content as it is discovered. It supports advanced capabilities such as headless browser rendering, background crawling tasks, and configurable rules that control crawl depth or ignored paths. These capabilities make the project suitable for building search indexers, data extraction pipelines, & SEO analysis tools.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Scraper of Death
    Scraper of Death is a web scraper. Multiple Scraping Methods Requests + BeautifulSoup (fast, lightweight) Selenium (JavaScript support, dynamic content)
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    SiteOne Crawler (desktop app)

    SiteOne Crawler (desktop app)

    A free, feature-rich web analyzer and exporter/cloner you will love!

    A free in-depth website analyzer providing audits of security, performance, SEO, accessibility and other technical aspects. Available as a desktop application for Windows/macOS/Linux and as a CLI tool for advanced users and CI/CD processes. It also includes an offline web page exporter (website clone, mirror).
    Downloads: 6 This Week
    Last Update:
    See Project
  • Fully Managed MySQL, PostgreSQL, and SQL Server Icon
    Fully Managed MySQL, PostgreSQL, and SQL Server

    Automatic backups, patching, replication, and failover. Focus on your app, not your database.

    Cloud SQL handles your database ops end to end, so you can focus on your app.
    Try Free
  • 10
    Catbird Linux

    Catbird Linux

    Linux for content creation, web scraping, coding, and data analysis.

    ...It is ready for work in Python, Lua, and Go languages, with numerous packages for web scraping or downloading data via API calls. Using Catbird Linux, it is possible to accomplish in depth stock market analysis, track weather trends, follow social media sentiment, or do other tasks in data science. The system is programmer friendly, ready for creating and running the tools you use to measure and understand your world. In addition to search and GPT tools, you have what you need to take notes, write reports or presentations, record and edit audio or video. ...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 11
    Laravel WebSockets

    Laravel WebSockets

    Websockets for Laravel

    Laravel WebSockets is a package for Laravel 5.7 and up that will get your application started with WebSockets in no time. It has a drop-in Pusher API replacement, a debug dashboard, real-time statistics and even allows you to create custom WebSocket controllers.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12

    S-Web42

    A Camel approach to website building

    Generate an output website tree from an input tree of arbitrary depth consisting of a set of pretty simple textfiles with a mixture of PIs, (uninterpreted) (X)HTML markup and normal text. Per-directory hooks may exist, perl(1) code may be embedded, and a sftp(1) batchfile will be automatically produced. Please do read README for more.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Go Safe Web

    Go Safe Web

    Secure-by-default HTTP servers in Go

    ...It also offers utilities for CSRF protection, secure cookies, and safe resource embedding that work well with Go’s standard library. By turning security posture into a first-class concern, go-safeweb helps teams achieve defense-in-depth without scattering security logic across a codebase.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Ultra

    Ultra

    A Leiningen plugin for a superior development environment

    Ultra is a Leiningen plugin for an absolutely kick-ass development environment. I've written a blog post describing Ultra in greater depth here. Ultra is the rare piece of essentially "finished" software. I've ironed out most of the bugs, and don't plan on adding much in the future. In other words: it's not unmaintained, it's just done. At the moment, Ultra doesn't have ClojureScript support at the REPL. The relevant upstream issue to track work on this is greglook/puget#27; from there, Whidbey will need to be updated, and then Ultra will be able to consume the changes. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    crawler4j

    crawler4j

    Open source web crawler for Java

    crawler4j is an open source web crawler for Java which provides a simple interface for crawling the Web. Using it, you can setup a multi-threaded web crawler in few minutes. You need to create a crawler class that extends WebCrawler. This class decides which URLs should be crawled and handles the downloaded page. shouldVisit function decides whether the given URL should be crawled or not. In the above example, this example is not allowing .css, .js and media files and only allows pages...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    lightcrawler

    lightcrawler

    Website crawler that audits site pages automatically with Lighthouse

    ...Lightcrawler supports configuration through a JSON configuration file, enabling users to customize how the crawler operates and which Lighthouse audits should be executed. Settings such as crawl depth and the number of concurrent browser instances can be configured to control how aggressively the crawler scans a site. It was created as a developer utility to help identify issues across an entire website more efficiently.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 17

    DrupalCSVFields

    Create Drupal 7 fields from CSV

    ...See https://sourceforge.net/p/drupalcsvfields/code/HEAD/tree/trunk/examples/basictest/ for an usage example. It worked for me, but use it at your own risk. I have not tested it in depth. Please backup your data before enabling the modules produced by the utility.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    beanbun

    beanbun

    PHP multiprocess web crawling framework with distributed support

    Beanbun is a web crawling framework written in PHP that is designed to help developers build scalable and customizable web scrapers. Beanbun focuses on simplicity while providing a flexible architecture that supports complex crawling workflows. It supports both normal execution mode and daemon mode, allowing crawlers to run continuously in long-running background processes when needed. Beanbun uses a downloader component based on Guzzle and integrates with Workerman to enable multi-process...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    SSEP - Site Search Engine PHP-Ajax

    SSEP - Site Search Engine PHP-Ajax

    A Free site search engine script build with PHP and Ajax.

    A Site Search engine script that uses MySQL to store your website's indexed pages, to add Search Functionality to Your Web Site. It is build with PHP and JavaScript, the search results are loaded via Ajax. The search system combine MySQL full text with SQL regexp, and words weight according to their location in the HTML elements, to determine the relevance of the search results. It can be included in any web site.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    Symmetric Sprite Editor

    Symmetric Sprite Editor

    Create symmetric sprites: browse, save to PC, or store online

    This was a multi-user sprite creation tool.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    Linkcrawler

    Linkcrawler

    Capable to "Crawl" a site and return a report of all links from it

    Java Desktop application capable to "Crawl" a site and return a report of the status of all the link present at the page, then it moves to another internal page and so on. LinkCrawlers provides a nice HTML5 report with the information of all link per WebPage, Easy to Read. This tool is useful for Web QA testers
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    ChanDown - 4chan Image Downloader

    ChanDown - 4chan Image Downloader

    Auto Rescanning - Search Terms - Regularly Updated With New Features

    ...This downloader packs loads of great features such as the search ability. Check the features section and be sure to let me know if you want a feature added. Coming Soon: - Wiki, explaining in depth how to use it more quickly (although its already pretty simple to use) - Ability to download the whole thread, not just images - Better multithreading - Ability to use proxies - Sort images download from searches into folders - Keep original image names - More responsive gui Be sure to let me know if you want any other features.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Heartbleed

    Heartbleed

    A checker (site and tool) for CVE-2014-0160

    ...The repo also serves as a historical snapshot of a watershed moment in modern Internet security and the practical lessons it forced the industry to learn. Developers and security engineers alike can study it to better appreciate protocol hardening, input validation, and defense-in-depth strategies.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    Pye

    Pye

    Pye - Javascript Engine For Drawing Fast, Fluid and Dynamic Pie Charts

    Pye is a JavaScript engine for drawing beautiful pie charts that are fast, fluid and dynamic and can be easily modified dynamically. As great new web technologies progress, it is handy to have a handy script or utility to do the job efficiently and quickly.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Regexp Automatic Downloader

    Regexp Automatic Downloader

    Download files were the links list point to a download page

    This is a simple program, that allows you to download several files from a webpage, using URL pattern matching, and also allows for you to open the page and then select either the file download link or another page from that page. This makes mass download of file from website easier.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • Next
MongoDB Logo MongoDB