[prev in list] [next in list] [prev in thread] [next in thread]
List: lucene-user
Subject: Preventing duplicate document insertion during optimize
From: "Kevin A. Burton" <burton () newsmonster ! org>
Date: 2004-04-30 23:47:19
Message-ID: 4092E587.10704 () newsmonster ! org
[Download RAW message or body]
Let's say you have two indexes each with the same document literal. All
the fields hash the same and the document is a binary duplicate of a
different document in the second index.
What happens when you do a merge to create a 3rd index from the first
two? I assume you now have two documents that are identical in one
index. Is there any way to prevent this?
It would be nice to figure out if there's a way to flag a field as a
primary key so that if it has already added it to just skip.
Kevin
--
Please reply using PGP.
http://peerfear.org/pubkey.asc
NewsMonster - http://www.newsmonster.org/
Kevin A. Burton, Location - San Francisco, CA, Cell - 415.595.9965
AIM/YIM - sfburtonator, Web - http://peerfear.org/
GPG fingerprint: 5FB2 F3E2 760E 70A8 6174 D393 E84D 8D04 99F1 4412
IRC - freenode.net #infoanarchy | #p2p-hackers | #newsmonster
["signature.asc" (application/pgp-signature)]
[prev in list] [next in list] [prev in thread] [next in thread]
Configure |
About |
News |
Add a list |
Sponsored by KoreLogic