Showing 3 open source projects for "algorithm"

View related business solutions
  • Our Free Plans just got better! | Auth0 Icon
    Our Free Plans just got better! | Auth0

    With up to 25k MAUs and unlimited Okta connections, our Free Plan lets you focus on what you do best—building great apps.

    You asked, we delivered! Auth0 is excited to expand our Free and Paid plans to include more options so you can focus on building, deploying, and scaling applications without having to worry about your security. Auth0 now, thank yourself later.
    Try free now
  • Fully managed relational database service for MySQL, PostgreSQL, and SQL Server Icon
    Fully managed relational database service for MySQL, PostgreSQL, and SQL Server

    Focus on your application, and leave the database to us

    Cloud SQL manages your databases so you don't have to, so your business can run without disruption. It automates all your backups, replication, patches, encryption, and storage capacity increases to give your applications the reliability, scalability, and security they need.
    Try for free
  • 1
    Genetic Algorithm File Fitter, GAFFitter for short, is a tool based on a genetic algorithm (GA) that tries to fit a collection of items, such as files/directories, into as few as possible volumes of a specific size (e.g. CDs or DVDs).
    Downloads: 24 This Week
    Last Update:
    See Project
  • 2
    Duplicate Files Finder is a cross-platform application for finding and removing duplicate files by deleting, creating hardlinks or creating symbolic links. A special algorithm minimizes the amount of data read from disk, so the program is very fast.
    Downloads: 41 This Week
    Last Update:
    See Project
  • 3

    qlsdup

    GUI based lightweight duplicate file finder / remover.

    ...This is basically a reimplementation of dupfinder as: it doesn't compile on my computer, the executable won't work either, the projects seems dead and I didn't wanna work through the original code. So Look'n'Feel should be somewhat the same, though some changes have been made. Algorithm is simple and suitable for large file sets with few differences: basic set of duplicate candidates is determined by file size, then candidates are compared byte-wise. So for files which don't start to differ by growing at the tail it reduces read operations greatly compared to hash-based comparision (i.e. test case: 40GB of files, 16GB RAM, many differences: initial compare (uncached by OS) around 300sec, subsequent compare (relevant file parts cached by OS) ca. 2sec) Note: this program is not related to lsdup
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next