Am Sonntag, den 29.05.2005, 12:02 +0200 schrieb Raphaël Slinckx:
> On dim, 2005-05-29 at 03:00 +0200, A. Pagaltzis wrote:
> > * A. Pagaltzis <pagaltzis@...> [2005-05-29 02:45]:
> > What I mean is, of course, if an entry in one feed has the same ID
> > as an entry in another, but the user has not declared that he
> > trusts one feed to be republishing entries from the other, then
> > you should be paranoid.
> As you said there is no reliable way to detect this, so i think liferea
> should not do anything to try to detect it, it will be hackish at more.
> This particular case should not interfere on the descussion on how to
> display multiple similar feeds.
This is correct, but both ideas share a similar implementation
with in my opininion quite some effort and significant runtime
overhead to match newly downloaded items against all cached
items. Of course this will be faster with a DB backend, but
even if it would take parts seconds for each item it sums up
when updating for example 5 feeds resulting in 30 new items.
Also this could cause pretty CPU intense operations.
I think such specialised data-mining features are have a bad
development+runtime effort versus usefulness ratio.
> > > In that case, if it’s a republisher, and the entry you received
> > > differs from the one you got from the original source, then you
> > > can refresh the original feed and check what that looks like.
> > And by this I mean:
> > * feed f1 has entry e1 with ID xyz
> > * feed f2 has an entry e2 different from e1, but also with ID xyz
> What i suggest to detect same unique id (that's weird :) ) on different
> posts, is to also take account of the title of the post. These two
> should be enough to determine if two items are identical or not. Maybe
> even the first of last characters.
> For the UI, i don't want to see there are multiple same item, otherwise
> it's just like it is now. So i suggest in the unread vFolder only one
> item appear maybe with an icon indicating that it is has duplicates, or
> maybe with an arrow so you can expand and see the duplicate items.
I fear I might disappoint you now, but my experience
is that you cannot maintain this feature complexity level
with two developers currently spending about 1 hour a day.
My mantra: the feature set and complexity must match the