The stuff here has no documentation and some of it may never be completed. This is my playground, use at your own risk.
ezMPEG is an easy-to-use and easy-to-understand MPEG1 video encoder API
lease-parser is a simple daemon that records the lease state changes of an ISC DHCP server to a database for historical reference. The data can be searched via a web search form that is provided with the tool.
A distributed highly customizable web search system designed to be able to include custem parsers to add additional searchable metadata from the content of a site as well as from the url of both the site and the referrer.
When released, FilmSearch will let you gain a huge amount of time: you'll no more have to scan every day the program of some dozens of TV-channels, just to find once per month something interesting enough to turn on the TV. Each user will be able t
A collection of software to implement search engine technology. The overall search technology is built on the individual components of this project, each component is released under the BSD License, and is written in the language most suited to its task.
A cross-platform, highly configurable, distributed web crawling system with crawlers optimized for speed.
OpenAnonymity consists of a module for apache 2.0 Webserver and a framework that enables you to control search engine spider indexing on a word level, contrary to on file level as in Robots exclusion. OA could force Spiders to follow this rules.
The goal of OpenParentalControls is to provide a user-contributed database of website age ratings, as well as a series of extensions for popular web browsers to honor, update and vote on these ratings.
PHP Wrapper Class For ht://Dig is a class I developed while desperately searching for something with similar capabilites. This class is intended to be much more thorough allowing for easily changing headers, footers, and templates. htdig + PHP = htPHP
This will be a implementation of Google PageRank Algorithm. For calculate the PR, could be use various PCs for speed up.
The SiCrawler (or Sensitive Information Crawler) is a web crawler designed to extract user defined sensitive information from web sites. This could be credit card/social security numbers, or a host of other information defined by regexes and plug-ins.
The Species Analyst (TSA) is a research project developing standards and software tools for access to the worlds natural history collection and observation databases.
The project is to use the search best practices available with various search engines and build an abstraction layer above the search engines for efficient and very relevant internet content search organized in user specified format.
ht://Check is more than a link checker. It's particularly suitable for checking broken links, anchors and web accessibility barriers, but retrieved data can also be used for Web structure mining. Uses a MySQL backend. Derived from ht://Dig.
iXDCC is an IRC bot for channel owners who want index the xdcc lists from the xdcc bots of their channels. It can also respond to @find commands, generate an xml "database" of packs and upload it through FTP on a remote host.
Utilities and C/C++-library API for mapping between IP-addresses and locations (ISO 3166 country codes and names).
lemonade - Indonesian Search Engine scripts. It purpose is to be one of the best search engine around in Indonesia, yet also for you to use it on your website at no cost. Will be available in both English and Indonesian Language.