Conference item icon

Conference item

Quantifying robustness of trust systems against collusive unfair rating attacks using information theory

Abstract:
Unfair rating attacks happen in existing trust and reputation systems, lowering the quality of the systems. There exists a formal model that measures the maximum impact of independent attackers [Wang et al., 2015] — based on information theory. We improve on these results in multiple ways: (1) we alter the methodology to be able to reason about colluding attackers as well, and (2) we extend the method to be able to measure the strength of any attacks (rather than just the strongest attack). Using (1), we identify the strongest collusion attacks, helping construct robust trust system. Using (2), we identify the strength of (classes of) attacks that we found in the literature. Based on this, we help to overcome a shortcoming of current research into collusion-resistance — specific (types of) attacks are used in simulations, disallowing direct comparisons between analyses of systems.
Publication status:
Published
Peer review status:
Peer reviewed

Actions


Access Document


Files:

Authors


More by this author
Institution:
University of Oxford
Division:
MPLS
Department:
Computer Science
Role:
Author


Publisher:
International Joint Conferences on Artificial Intelligence
Host title:
Proceedings of the 24th International Joint Conference on Artificial Intelligence
Journal:
Proceedings of the 24th International Joint Conference on Artificial Intelligence More from this journal
Pages:
111-117
Publication date:
2015-09-25
Acceptance date:
2015-04-16
ISBN:
9781577357384


Keywords:
Pubs id:
pubs:891856
UUID:
uuid:c302e02d-e13b-4481-a40e-f5223d7e5101
Local pid:
pubs:891856
Source identifiers:
891856
Deposit date:
2018-07-31

Terms of use



Views and Downloads






If you are the owner of this record, you can report an update to it here: Report update to this record

TO TOP