Showing 2 open source projects for "git:/git.code.sf.net/p/docfetcher/code"

View related business solutions
  • Gen AI apps are built with MongoDB Atlas Icon
    Gen AI apps are built with MongoDB Atlas

    Build gen AI apps with an all-in-one modern database: MongoDB Atlas

    MongoDB Atlas provides built-in vector search and a flexible document model so developers can build, scale, and run gen AI apps without stitching together multiple databases. From LLM integration to semantic search, Atlas simplifies your AI architecture—and it’s free to get started.
    Start Free
  • Keep company data safe with Chrome Enterprise Icon
    Keep company data safe with Chrome Enterprise

    Protect your business with AI policies and data loss prevention in the browser

    Make AI work your way with Chrome Enterprise. Block unapproved sites and set custom data controls that align with your company's policies.
    Download Chrome
  • 1
    Text Line Duplicate Remover

    Text Line Duplicate Remover

    Remove duplicate lines from your text

    This standalone offline web browser tool helps you remove duplicate lines from your text, with additional text processing options. Simply open it in your browser by double-clicking the html file. It also includes the source code too. I made this when I was working with long lists of entries and needed something to automatically clean them up. As a bonus you can also change the Sentence Case of the text, make it lowercase, UPPERCASE or Sentence case.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 2

    qlsdup

    GUI based lightweight duplicate file finder / remover.

    ...This is basically a reimplementation of dupfinder as: it doesn't compile on my computer, the executable won't work either, the projects seems dead and I didn't wanna work through the original code. So Look'n'Feel should be somewhat the same, though some changes have been made. Algorithm is simple and suitable for large file sets with few differences: basic set of duplicate candidates is determined by file size, then candidates are compared byte-wise. So for files which don't start to differ by growing at the tail it reduces read operations greatly compared to hash-based comparision (i.e. test case: 40GB of files, 16GB RAM, many differences: initial compare (uncached by OS) around 300sec, subsequent compare (relevant file parts cached by OS) ca. 2sec) Note: this program is not related to lsdup
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next