Full access to Enterprise features. No credit card required.
What if You Could Automate 90% of Your Repetitive Tasks in Under 30 Days? At ServoDesk, we help businesses like yours automate operations with AI, allowing you to cut service times in half and increase productivity by 25% - without hiring more staff.
Try ServoDesk for free
Keep company data safe with Chrome Enterprise
Protect your business with AI policies and data loss prevention in the browser
Make AI work your way with Chrome Enterprise. Block unapproved sites and set custom data controls that align with your company's policies.
Simple SQL Client for lightweight data analysis. You can share the result with gist. Supports MySQL, PostgreSQL (Amazon Redshift), SQLite3, Google BigQuery, Treasure Data, Amazon Athena. You can download and install from Web Site or Releases.
WallPaper (alias crawlpaper) is a desktop changer (NOT a screensaver) which includes a webcrawler for picture download, an audio stream ripper, an audio player, a mini mp3 tag editor,etc. Also included support for .zip and .rar files and an interface to the BerkleyDB code for small databases.
PHP Search is a search engine script that searches a MySQL database for links and descriptions much like google. Manual adding of Data. Crawler Coming soon! Demo at http://www.jhosting.tk/admin/search/search.php
Axe Credit Portal - ACP- is axefinance’s future-proof AI-driven solution to digitalize the loan process from KYC to servicing, available as a locally hosted or cloud-based software.
Banks, lending institutions
Founded in 2004, axefinance is a global market-leading software provider focused on credit risk automation for lenders looking to provide an efficient, competitive, and seamless omnichannel financing journey for all client segments (FI, Retail, Commercial, and Corporate.)
Decima is a database that was designed to support time-series data mining. It consists of PostgreSQL custom type definition, implementation of GiST index for that type and snowflake database schema.
Larbin is a Webcrawler intended to fetch a large number of Web pages, it should be able to fetch more than 100 millions pages on a standard PC with much u/d. This set of PHP and Perl scripts, called webtools4larbin, can handle the output of Larbin and p
A VB Webcrawler that is currently under construction with the goal to be able to crawl and index the net most likely by distributed computing (via network).
Web Textual eXtraction Tools C++ Parallel webcrawler, noun phrase idenification, Multi-lingual Part of Speech Tagging, Tarjan's Algorithm, Co-RelationShip Mappings...