diff --git a/router/java/src/net/i2p/router/networkdb/kademlia/SearchJob.java b/router/java/src/net/i2p/router/networkdb/kademlia/SearchJob.java index f49ca4276a82db02738f1354aa1eaea669fc1cb7..8b7c97e0ad53fef2ff6e97672d3b9556a99b4ccd 100644 --- a/router/java/src/net/i2p/router/networkdb/kademlia/SearchJob.java +++ b/router/java/src/net/i2p/router/networkdb/kademlia/SearchJob.java @@ -841,15 +841,16 @@ class SearchReplyJob extends JobImpl { // are next in our search path... // note: no need to think about shitlisted targets in the netdb search, given // the floodfill's behavior - //if (getContext().shitlist().isShitlisted(peer)) { + // This keeps us from continually chasing blocklisted floodfills + if (getContext().shitlist().isShitlisted(peer)) { // if (_log.shouldLog(Log.INFO)) // _log.info("Not looking for a shitlisted peer..."); // getContext().statManager().addRateData("netDb.searchReplyValidationSkipped", 1, 0); - //} else { + } else { //getContext().netDb().lookupRouterInfo(peer, new ReplyVerifiedJob(getContext(), peer), new ReplyNotVerifiedJob(getContext(), peer), _timeoutMs); //_repliesPendingVerification++; shouldAdd = true; - //} + } } else { if (_log.shouldLog(Log.INFO)) _log.info("Peer " + _peer.toBase64() + " sends us bad replies, so not verifying " + peer.toBase64());