Showing 291 open source projects for "anpr using python"

View related business solutions
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    More flexibility. More control.

    Generate interest, access liquidity without selling, and execute trades seamlessly. All in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 1
    ruia

    ruia

    Async Python framework for fast and flexible web scraping spiders

    Ruia is an asynchronous web scraping micro-framework built for Python that focuses on simplicity, speed, and flexibility when creating web crawlers. Ruia is powered by Python’s asyncio library along with aiohttp, enabling developers to perform concurrent network requests efficiently and scrape data from websites with minimal overhead. Ruia follows a “write less, run faster” philosophy, emphasizing concise code and streamlined spider development. It provides a structured approach to building...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 2
    AWS IoT Greengrass Core SDK

    AWS IoT Greengrass Core SDK

    SDK to use with functions running on Greengrass Core using Python

    The AWS IoT Greengrass Core SDK is meant to be used by AWS Lambda functions running on an AWS IoT Greengrass Core. It will enable Lambda functions to invoke other Lambda functions deployed to the Greengrass Core, publish messages to the Greengrass Core and work with the local Shadow service. To use the AWS IoT Greengrass Core SDK, you must first import the AWS IoT Greengrass Core SDK in your Lambda function as you would with any other external libraries. You then need to create a client for...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    GoogleScraper

    GoogleScraper

    Python tool for scraping search engine results from many providers

    GoogleScraper is a Python-based tool designed to automatically collect and process search engine results from multiple providers. It enables developers and researchers to programmatically query search engines and extract useful information such as links, titles, and result descriptions. GoogleScraper supports several major search engines and can be used to gather structured datasets from search result pages for further analysis. It provides two different scraping approaches: sending direct...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    TCellXTalk

    TCellXTalk

    TCellXTalk Web-App from LP CSIC/UAB

    TCellXTalk is a comprehensive database of experimentally detected phosphorylation, ubiquitination and acetylation sites in human T cells. The web-app at www.TCellXTalk.org makes TCellXTalk accessible from Internet, and enables the in silico prediction of potential co-modified peptides to facilitate their experimental detection, using targeted or directed mass spectrometry, for the study of protein post-translational modification cross-talk. More detailed information on TCellXTalk and...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Train ML Models With SQL You Already Know Icon
    Train ML Models With SQL You Already Know

    BigQuery automates data prep, analysis, and predictions with built-in AI assistance.

    Build and deploy ML models using familiar SQL. Automate data prep with built-in Gemini. Query 1 TB and store 10 GB free monthly.
    Try Free
  • 5
    LymPHOS2

    LymPHOS2

    LymPHOS2 Web-App

    LymPHOS2 is a web-based Application at www.LymPHOS.org containing peptidic and protein sequences and spectrometric information on the PhosphoProteome of human T-Lymphocytes. - Nguyen, TD., Vidal-Cortes, O., Gallardo, Ó., Abian, J., Carrascal, M., LymPHOS 2.0: an update of a phosphosite database of primary human T cells. Database 2015, 2015. DOI: 10.1093/database/bav115 - Carrascal, M., Ovelleiro, D., Casas, V., Gay, M., Abian, J., Phosphorylation analysis of primary human T lymphocytes...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    xsrfprobe

    xsrfprobe

    Advanced toolkit for detecting and exploiting CSRF vulnerabilities

    XSRFProbe is an advanced security auditing toolkit designed to detect and analyze Cross Site Request Forgery (CSRF/XSRF) vulnerabilities in web applications. It uses an automated crawling engine that continuously scans a target application, collects forms and endpoints, and evaluates them for potential CSRF weaknesses. XSRFProbe performs numerous systematic checks to determine whether a web endpoint is vulnerable, including inspection of anti-CSRF tokens, cookie validation behavior, and...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 7
    Backuso

    Backuso

    Backup software for your data to local and remote SSH systems

    ...This could be a NAS (Network Attached Storage) in your local LAN, or a server share on the internet. Backuso is configured with a simple text file. You can start the backups manually or automatically, e.g. using crontab. Backuso is based on Python 3 and uses the software rsync as its core element. Hence it runs on systems where rsync is available (Linux, FreeBSD, Apple).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    v2rayL

    v2rayL

    v2ray linux GUI

    V2Ray is a tool under Project V. Project V includes a series of tools to help you create your own customized network system. And V2Ray belongs to the core one. Simply put, V2Ray is a proxy software similar to Shadowsocks, but has more advantages than Shadowsocks.v2ray linux client, using pyqt5 to write GUI interface, the core is based on v2ray-core (v2ray-linux-64) vmess supports websocket, mKcp, and tcp. There may be some bugs in the current program, but they have not been tested. If you...
    Downloads: 14 This Week
    Last Update:
    See Project
  • 9

    pyindi-client

    Python binding to the libindi library

    This project offers a python binding to the libindi library. It uses the Simplified Wrapper and Interface Generator (SWIG). You may generate a PyIndi module which mainly defines an IndiClient class. This class could be used to build python scripts able to interact with indi servers using sendNew* methods and implementing new* methods of the BaseMediator class.
    Downloads: 1 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 10
    AET

    AET

    Detects visual changes on websites and performs page health checks

    AET is a system that detects visual changes on websites and performs basic page health checks (like w3c compliance, accessibility, HTTP status codes, JS Error checks and others). AET is designed as a flexible system that can be adapted and tailored to the regression requirements of a given project. The tool has been developed to aid front-end client-side layout regression testing of websites or portfolios, in essence assessing the impact or change of a website from one snapshot to the next.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 11
    Requests-HTML

    Requests-HTML

    Pythonic HTML Parsing for Humans

    This library intends to make parsing HTML (e.g. scraping the web) as simple and intuitive as possible. When using this library you automatically get full JavaScript support! (Using Chromium, thanks to puppeteer) CSS Selectors (a.k.a jQuery-style, thanks to PyQuery). XPath Selectors, for the faint of heart. Mocked user-agent (like a real web browser). Automatic following of redirects. Connection–pooling and cookie persistence. The Requests experience you know and love, with magical parsing...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Cloud Filer

    Cloud Filer

    S3 file management tool with encryption

    Cloud Filer is a graphical cloud file management tool and S3 client. Cloud Filer allows you to upload and download files to cloud storage and can be used as part of your backup strategy. Cloud Filer includes pre-internet encryption and file name obfuscation making access to your cloud data useless without the right password. Download and install using pip install cloudfiler - see also https://pypi.org/project/cloudfiler/
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Jupyter Server Proxy

    Jupyter Server Proxy

    Jupyter notebook server extension to proxy web services.

    Jupyter Server Proxy lets you run arbitrary external processes (such as RStudio, Shiny Server, Syncthing, PostgreSQL, Code Server, etc) alongside your notebook server and provide authenticated web access to them using a path like /rstudio next to others like /lab. Alongside the Python package that provides the main functionality, the JupyterLab extension (@jupyterhub/jupyter-server-proxy) provides buttons in the JupyterLab launcher window to get to RStudio for example.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Rendora

    Rendora

    dynamic server-side rendering using headless Chrome

    Rendora is a dynamic renderer to provide zero-configuration server-side rendering mainly to web crawlers in order to effortlessly improve SEO for websites developed in modern Javascript frameworks such as React.js, Vue.js, Angular.js, etc. Rendora works totally independently of your frontend and backend stacks. Rendora can be seen as a reverse HTTP proxy server sitting between your backend server (e.g. Node.js/Express.js, Python/Django, etc...) and potentially your frontend proxy server...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Transcrypt

    Transcrypt

    Python in the Browser

    Lean and mean Python 3.6 to JavaScript compiler. Supports multiple inheritance, operator overloading and Python source level debugging, even of minified Javascript files. Transcrypt code is as fast and compact as its Javascript counterpart, and it is precompiled for page load speed. You can now develop your web applications completely in Python, with full access to any Javascript library.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    mzitu

    mzitu

    Python crawler that downloads image galleries and analyzes titles

    Using text segmentation and frequency analysis, the project can create a word cloud representing common keywords found in the dataset. This makes the repository both a scraping example and a small data analysis experiment built around the collected content. Overall, mzitu serves as a learning-oriented implementation of Python web scraping, data processing, and visualization techniques.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 17
    Twitter Intelligence

    Twitter Intelligence

    Twitter Intelligence OSINT project performs tracking and analysis

    A project written in Python for Twitter tracking and analysis without using Twitter API. This project is a Python 3.x application. The package dependencies are in the file requirements.txt. Run that command to install the dependencies. SQLite is used as the database. Tweet data is stored on the Tweet, User, Location, Hashtag, HashtagTweet tables. The database is created automatically. analysis.py performs analysis processing.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18

    blog99

    A blog engine that does html and gopher

    This is the blog engine for HTML and Gopher. Blog entries are written as html files. For HTML, it is an Apache/MySQL/Python application using WSGI. For Gopher, it is Gophernicus/MySQL/Python using CGI.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19

    pyUrlMin

    Ultra-minimalist Python/Mongo URL Shortner

    With as little code as possible, this micro-app creates shortened URLs, stores them in MongoDB using the Flask framework. Currently in use on https://findmate.app and http://urltools.io/ HOWEVER: This should not really be used in a high-volume production environment and should really be used for educational purposes only. I assume no liability for mis-use of this app, use it at your own risk. There are potentially security issues, so be forewarned that this is just for TESTING only.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    WeChatSogou

    WeChatSogou

    Python library to crawl and retrieve data from WeChat accounts

    WechatSogou is an open source Python library designed to retrieve data from WeChat official accounts by using the Sogou WeChat search service as its data source. It provides developers with a programmatic way to search for public accounts and collect article information without manually browsing the search interface. It functions as a crawler interface that sends requests to the search engine, retrieves results, and converts the returned pages into structured data that can be used in applications or analysis pipelines. ...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 21
    pyspider

    pyspider

    A powerful Spider(Web Crawler) system in Python

    pyspider is a powerful Spider(Web Crawler) system in Python. Components are connected by message queue. Every component, including message queue, is running in their own process/thread, and replaceable. That means, when process is slow, you can have many instances of processor and make full use of multiple CPUs, or deploy to multiple machines. This architecture makes pyspider really fast. benchmarking.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    gain

    gain

    Asyncio-based Python framework for building fast web crawling spiders

    Gain is a Python web crawling framework designed to simplify the process of building efficient and scalable web scrapers. It is built on top of asynchronous technologies such as asyncio, aiohttp, and uvloop to support high-performance crawling with concurrent network requests. It provides a structured framework for creating spiders that can navigate websites, extract structured data, and process the collected results.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 23
    haipproxy

    haipproxy

    Distributed proxy IP pool for web crawlers using Scrapy and Redis

    ...It automatically crawls proxy resources from the internet and aggregates them into a centralized pool that can be accessed by distributed spiders and scraping systems. It is built using Python and relies on Scrapy for high-performance crawling while Redis is used for data storage, communication, and task coordination between components. It includes crawlers that discover proxy servers, validators that test proxy availability and performance, and schedulers that manage crawling and validation tasks. HAipproxy aims to maintain a high availability proxy pool with low latency so that scraping frameworks can rotate proxies efficiently and avoid blocking during large-scale data collection. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    Toapi

    Toapi

    Convert websites into structured APIs automatically with Python tool

    Toapi is a Python library designed to transform ordinary websites into usable API services. Instead of building a traditional web crawler that collects and stores data before exposing it through an API, Toapi simplifies the process by allowing developers to define data structures that automatically generate an API layer from existing web pages. It works by parsing HTML content from a source site and mapping selected elements into structured data that can be returned as JSON through API...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    uncaptcha

    uncaptcha

    Defeating Google's audio reCaptcha with 85% accuracy

    uncaptcha is an open-source proof-of-concept system designed to demonstrate vulnerabilities in Google’s audio reCAPTCHA challenges by automatically solving them using speech recognition techniques. The project uses browser automation to navigate to CAPTCHA challenges, extract audio files, and process them through multiple speech-to-text services. By combining outputs from several transcription engines, the system increases the likelihood of correctly identifying the spoken digits or phrases...
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB