As a hobby, I am writing a simple and primitive distributed web search engine. It occurred to me that it currently has no protection against malicious peers trying to skew the search results.
The principle idea of the project is:
- To store the inverse indices and ranking factors in a Kad (Kademlia) DHT (Distributed Hash Table).
- The peers update the inverse indices as they crawl web.
I used Google Scholar to find a solution. It seems that most of the authors of proposed P2P web search ignores the above-mentioned problem.
I think I need some kind of reputation system or trust metrics. My knowledge in this domain is sufficiently lacking. I would very much appreciate a few pointers.