Showing 2 open source projects for "git:/git.code.sf.net/p/docfetcher/code"

View related business solutions
  • Level Up Your Cyber Defense with External Threat Management Icon
    Level Up Your Cyber Defense with External Threat Management

    See every risk before it hits. From exposed data to dark web chatter. All in one unified view.

    Move beyond alerts. Gain full visibility, context, and control over your external attack surface to stay ahead of every threat.
    Try for Free
  • Gen AI apps are built with MongoDB Atlas Icon
    Gen AI apps are built with MongoDB Atlas

    The database for AI-powered applications.

    MongoDB Atlas is the developer-friendly database used to build, scale, and run gen AI and LLM-powered apps—without needing a separate vector database. Atlas offers built-in vector search, global availability across 115+ regions, and flexible document modeling. Start building AI apps faster, all in one place.
    Start Free
  • 1
    Text Line Duplicate Remover

    Text Line Duplicate Remover

    Remove duplicate lines from your text

    This standalone offline web browser tool helps you remove duplicate lines from your text, with additional text processing options. Simply open it in your browser by double-clicking the html file. It also includes the source code too. I made this when I was working with long lists of entries and needed something to automatically clean them up. As a bonus you can also change the Sentence Case of the text, make it lowercase, UPPERCASE or Sentence case.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 2

    qlsdup

    GUI based lightweight duplicate file finder / remover.

    ...This is basically a reimplementation of dupfinder as: it doesn't compile on my computer, the executable won't work either, the projects seems dead and I didn't wanna work through the original code. So Look'n'Feel should be somewhat the same, though some changes have been made. Algorithm is simple and suitable for large file sets with few differences: basic set of duplicate candidates is determined by file size, then candidates are compared byte-wise. So for files which don't start to differ by growing at the tail it reduces read operations greatly compared to hash-based comparision (i.e. test case: 40GB of files, 16GB RAM, many differences: initial compare (uncached by OS) around 300sec, subsequent compare (relevant file parts cached by OS) ca. 2sec) Note: this program is not related to lsdup
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next