Showing 33 open source projects for "statistical process control"

View related business solutions
  • Fully Managed MySQL, PostgreSQL, and SQL Server Icon
    Fully Managed MySQL, PostgreSQL, and SQL Server

    Automatic backups, patching, replication, and failover. Focus on your app, not your database.

    Cloud SQL handles your database ops end to end, so you can focus on your app.
    Try Free
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 1
    Spider

    Spider

    High-performance Rust web crawler and scraper for large-scale data

    ...Spider can operate concurrently across many pages, allowing it to gather large datasets in a short period of time. Spider also provides mechanisms for subscribing to crawl events so developers can process page data such as URLs, status codes, or HTML content as it is discovered. It supports advanced capabilities such as headless browser rendering, background crawling tasks, and configurable rules that control crawl depth or ignored paths. These capabilities make the project suitable for building search indexers, data extraction pipelines, & SEO analysis tools.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 2
    MergeSentinel

    MergeSentinel

    Controls the approval of merge requests in GitLab Community Edition

    MergeSentinel is a Go application designed to enhance the merge request (MR) process in GitLab. It listens for HTTP calls from GitLab project webhooks when a merge request action occurs. Depending on the rules configured, it will enable or disable the button used to accept the merge request, ensuring that all predefined criteria are met before a merge can be approved.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Outline Server

    Outline Server

    Outline Manager, developed by Jigsaw

    ...However, with most VPN providers, you don't always know who controls the VPN and who has access to your data. Creating your own VPN server used to be a complicated process…until now. Outline allows anyone to create their own VPN, run it, and share access to it. Outline is designed to offer block resistance. It allows you to control your server settings, including its location. Furthermore, Outline is a 100% open-source project that uses transparent technology and has undergone audits by two security companies.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 4
    Etaoin

    Etaoin

    Pure Clojure Webdriver protocol implementation

    ...Use the Etaoin library to automate a browser, test your frontend behavior, simulate human actions or whatever you want. Currently supports Chrome, Firefox, Safari, and Edge. Can either connect to a remote WebDriver process or have Etaoin launch one for you. Run your unit tests directly from Emacs by pressing C-t t as usual. Can imitate human-like behavior (delays, typos, etc). Control a SpaceX ISS Docking Simulator. Automatically watch/refresh the browser when making edits to Babashka book. Generate contributor badges for rewrite-clj, cljdoc and test-doc-blocks via a doc-update-readme babashka task.
    Downloads: 0 This Week
    Last Update:
    See Project
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 5
    Geziyor

    Geziyor

    Blazing fast Go framework for web crawling and data scraping tasks

    Geziyor is a high-performance web crawling and web scraping framework built for the Go programming language. It is designed to help developers crawl websites and extract structured information from web pages efficiently. It focuses on speed and scalability, allowing large numbers of requests to be processed concurrently. Geziyor supports use cases such as data mining, monitoring web content, and automated testing workflows. It provides a flexible architecture where developers define parsing...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    MyBatis Pagination

    MyBatis Pagination

    A pagination plugin

    If you are using MyBatis, it is recommended to try this pagination plugin. This must be the most convenient pagination plugin. PageHelper supports any complex single-table, multi-table queries. As to some special cases, please refer to the Important notes. Through a comprehensive sample code and test, the basic usage of adding, deleting, modifying, and checking operations in the MyBatis XML mode and annotation mode is explained, and the application of dynamic SQL in different aspects and the...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    watercrawl

    watercrawl

    AI-ready web crawler that extracts and structures website content

    WaterCrawl is an open source web crawling and data extraction platform designed to transform website content into structured data suitable for machine learning and AI workflows. It enables developers and researchers to crawl web pages, extract meaningful information, and convert it into formats that are easier to process and analyze. It provides a modern crawling system that can automatically navigate links, control crawl depth, and collect content from targeted sections of a website. WaterCrawl supports customizable extraction rules so users can focus only on relevant elements while ignoring unnecessary page components. WaterCrawl also offers real-time monitoring capabilities, allowing users to track crawling progress, performance metrics, and errors during large data collection jobs. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    RPA for Python

    RPA for Python

    Python package for doing RPA

    Python package for doing RPA. RPA for Python's simple and powerful API makes robotic process automation fun! You can use it to quickly automate away repetitive time-consuming tasks on websites, desktop applications, or the command line. See sample Python script, the RPA Challenge solution, and RedMart groceries example. To send a Telegram app notification, simply look up @rpapybot to allow receiving messages. To automate Chrome browser invisibly, use headless mode. To run 10X faster instead...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Scrapyd

    Scrapyd

    A service daemon to run Scrapy spiders

    Scrapyd can manage multiple projects and each project can have multiple versions uploaded, but only the latest one will be used for launching new spiders. A common (and useful) convention to use for the version name is the revision number of the version control tool you’re using to track your Scrapy project code. For example: r23. The versions are not compared alphabetically but using a smarter algorithm (the same packaging uses) so r10 compares greater to r9, for example. Scrapyd is an application (typically run as a daemon) that listens to requests for spiders to run and spawns a process for each one. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 10
    mlscraper

    mlscraper

    ML-based HTML scraper that learns extraction rules from examples

    ...It analyzes those examples within the HTML document and determines patterns or rules that can be used to extract the same type of information from similar pages. Once trained, the generated scraper can process new pages and return the extracted data in structured formats such as dictionaries or lists. This approach simplifies web scraping tasks by shifting the focus from rule-writing to example-based training. Internally, the project processes HTML documents, identifies relevant elements in the DOM, and builds extraction logic based on statistical or heuristic analysis of the training samples. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Browser Pool

    Browser Pool

    A Node.js library to easily manage and rotate a pool of web browsers

    Browser Pool is a small, but powerful and extensible library, that allows you to seamlessly control multiple headless browsers at the same time with only a little configuration, and a single function call. Currently it supports Puppeteer, Playwright and it can be easily extended with plugins. We created Browser Pool because we regularly needed to execute tasks concurrently in many headless browsers and their pages, but we did not want to worry about launching browsers, closing browsers,...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    OpenDocMan

    OpenDocMan

    PHP Document Management System (DMS)

    OpenDocMan is a web based document management system (DMS) written in PHP designed to comply with ISO 17025 and OIE standard for document management. It features fine grained control of access to files, and automated install and upgrades.
    Downloads: 13 This Week
    Last Update:
    See Project
  • 13
    Abot

    Abot

    Fast and flexible C# framework for building customizable web crawlers

    Abot is an open source C# web crawler framework designed to help developers efficiently crawl and process web content. It focuses on speed, flexibility, and extensibility while handling the complex low-level tasks involved in web crawling. It manages essential components such as multithreading, HTTP requests, scheduling, and link parsing so developers can focus on processing the collected data. Abot follows a modular architecture that allows developers to customize nearly every stage of the crawl process by implementing or replacing core interfaces. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    gocrawl

    gocrawl

    Polite concurrent web crawler library for Go with flexible hooks

    gocrawl is a lightweight web crawling library written in the Go programming language that enables developers to build custom web crawlers and data extraction tools. gocrawl focuses on providing a minimal yet powerful crawling engine that can be easily extended and adapted for different web scraping or indexing tasks. It is designed to be polite when accessing websites by respecting crawling rules such as robots.txt policies and applying crawl delays for each host. It executes requests...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 15
    ruia

    ruia

    Async Python framework for fast and flexible web scraping spiders

    Ruia is an asynchronous web scraping micro-framework built for Python that focuses on simplicity, speed, and flexibility when creating web crawlers. Ruia is powered by Python’s asyncio library along with aiohttp, enabling developers to perform concurrent network requests efficiently and scrape data from websites with minimal overhead. Ruia follows a “write less, run faster” philosophy, emphasizing concise code and streamlined spider development. It provides a structured approach to building...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 16
    Muon SSH Terminal/SFTP client

    Muon SSH Terminal/SFTP client

    Graphical SFTP client and terminal emulator with helpful utilities

    ...This project is being renamed as previous name "Snowflake" is confusing since there is already a popular product with the same name. Muon is a graphical SSH client. It has an enhanced SFTP file browser, SSH terminal emulator, remote resource/process manager, server disk space analyzer, remote text editor, huge remote log viewer, and lots of other helpful tools, which makes it easy to work with remote servers. Muon provides functionality similar to web-based control panels but, it works over SSH from the local computer, hence no installation is required on the server. It runs on Linux and Windows. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 17
    Create Website From Text File

    Create Website From Text File

    A Website Builder / Creator Script to Create Website From a Text File

    Please watch the video below to see this website builder / creator script in action. in less than 5 Minutes, you can create a website / webpage, from a text file, using the script enclosed, in the zip file downloaded from above. A custom Puppy Linux Operating System has been created for this purpose of running the script on various client computers. You can download it here : https://sourceforge.net/projects/command-output-to-html-table/files/OS/ Wherever possible, give...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    gain

    gain

    Asyncio-based Python framework for building fast web crawling spiders

    Gain is a Python web crawling framework designed to simplify the process of building efficient and scalable web scrapers. It is built on top of asynchronous technologies such as asyncio, aiohttp, and uvloop to support high-performance crawling with concurrent network requests. It provides a structured framework for creating spiders that can navigate websites, extract structured data, and process the collected results.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 19
    Empire

    Empire

    A PaaS built on top of Amazon EC2 Container Service (ECS)

    Empire is a control layer on top of Amazon EC2 Container Service (ECS) that provides a Heroku-like workflow. It conforms to a subset of the Heroku Platform API, which means you can use the same tools and processes that you use with Heroku, but with all the power of EC2 and Docker. Empire is targeted at small to medium-sized startups that are running a large number of microservices and need more flexibility than what Heroku provides. You can read the original blog post about why we built...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    AWS Serverless Workshops

    AWS Serverless Workshops

    Code and walkthrough labs to set up serverless applications

    This repository contains a collection of workshops and other hands-on content that will guide you through building various serverless applications using AWS Lambda, Amazon API Gateway, Amazon DynamoDB, AWS Step Functions, Amazon Kinesis, and other services. This workshop shows you how to build a dynamic, serverless web application. You'll learn how to host static web resources with Amazon S3, how to use Amazon Cognito to manage users and authentication, and how to build a RESTful API for...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    Parallec

    Parallec

    Fast Parallel Async HTTP/SSH/TCP/UDP/Ping Client Java Library

    ...Parallec means Parallel Client (pronounced as "para-like"). Open Source from eBay Cloud. A convenient response context passes any object you need when handling a response. Process data anyway and send it anywhere. Intuitive builder pattern APIs make parallel requests exceedingly easy. Input target hosts from list, string, text, json path, and cms query from local or URLs. Auto progress polling to handle async APIs such as "download packages" or "create compute", enabling task-level concurrency control and orchestration. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 22
    SoyBeans

    SoyBeans

    A Task Processing Framework for Java

    Soy Beans is an HTTP request processing framework written in Java. Written as an alternative to frameworks like Struts and Stripes, it provides a robust and extremely flexible API enabling rapid deployment, and dynamic configuration.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23

    AWK Statistics

    awkStatistics generates statistics from csv-data with awk.

    AwkStatistics is a script which is used to generate statistical data for csv data by using awk. It takes two columns from the input file, one is used to identify the row(option -i) and the other one is the column where the values are taken from (option -v). The script was designed for performance engineers to be able to analyze CSV-Data on the fly and have a good possibility to automate the analysis process.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24

    SE Auditor

    Free SEO audit software.

    ...Unlike other programs of this type SE Auditor is: - free - fast, do not waste your time (the audit process goes smooth as butter) - simple to understand interface (you can see at a glance) - shows the data important in optimizing your site - based on the collected data allows to generate a PDF report with hints - is accurate (for example in counting words, characters, calculating text/html ratio..)
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    ...Core features are as follows: Database separate read and write speed memcached buffering speed HTML static pages acceleration Images watermarked image server separation image support CDN DNS round robin mirror server Multi-language support multi-currency support Orders report statistical analysis, product price concessions deal Page layout can be flexibly configured, multi-server load balanced layout Safety reliability, all the details of the optimization process Demo Reception: http://www.hecart.com/ Demo backstage account: demo Demo backstage Password: demo Demo Background: http://admin.hecart.com/
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • Next
MongoDB Logo MongoDB