Computational Trust and
Reputation Models
Andrew Diniz da Costa
[email protected]
Presentation Outline
• Part 1: Introduction
– Motivation
– Some definitions
• Part 2: Computational trust and reputation models
– eBay/OnSale
– SPORAS & HISTOS
– Fire Model
– Governance Framework
• Part 3: ART-Testbed
– Overview
Andrew Diniz da Costa © LES/PUC-Rio
2
Presentation Outline
• Part 1: Introduction
– Motivation
– Some definitions
• Part 2: Computational trust and reputation models
– eBay/OnSale
– SPORAS & HISTOS
– Fire Model
– Governance Framework
• Part 3: ART-Testbed
– Overview
Andrew Diniz da Costa © LES/PUC-Rio
3
What we are talking about ...
Andrew Diniz da Costa © LES/PUC-Rio
4
What we are talking about ...
Andrew Diniz da Costa © LES/PUC-Rio
5
Advantages of trust and reputation mechanisms
• Agents can obtain data from others agents.
• Shared experience.
• Decide on which to trust
Andrew Diniz da Costa © LES/PUC-Rio
6
Problems of trust and reputation mechanisms
• Not all kind of environments are suitable to apply these
mechanisms.
• Exclusion must be a punishment
• What is trust?
• What is reputation?
Andrew Diniz da Costa © LES/PUC-Rio
7
Trust
• Some statements we like:
• “Trust begins where knowledge ends: trust provides a basis
dealing with uncertain, complex, and threatening images of
the future.” [Luhmann,1979]
• “There are no obvious units in which trust can be
measured,” [Dasgupta, 2000]
Andrew Diniz da Costa © LES/PUC-Rio
8
Reputation
• Some definitions:
• “The estimation of the consistency over time of an attribute
or entity” [Herbig et al.]
• “Information that individuals receive about the behaviour of
their partners from third parties and that they use to decide
how to behave themselves” [Buskens, Coleman...]
• “The opinion others have of us”
Andrew Diniz da Costa © LES/PUC-Rio
9
What is a good trust model?
• A good trust model should be [Fullam et al, 05]:
• Accurate
– provide good previsions
• Adaptive
– evolve according to behaviour of others
• Multi-dimensional
– Consider different agent characteristics
• Efficient
– Compute in reasonable time and cost
Andrew Diniz da Costa © LES/PUC-Rio
10
Why using a trust model in a MAS ?
• Trust models allow:
– Identifying and isolating untrustworthy agents
– Evaluating an interaction’s utility
– Deciding whether and with whom to interact
Andrew Diniz da Costa © LES/PUC-Rio
11
Presentation Outline
• Part 1: Introduction
– Motivation
– Some definitions
• Part 2: Computational trust and reputation models
– eBay/OnSale
– SPORAS & HISTOS
– Fire model
– Governance Framework
• Part 3: ART-Testbed
– Overview
Andrew Diniz da Costa © LES/PUC-Rio
12
eBay model
• Context: e-commerce
– Model oriented to support trust between buyer and seller
– Buyer has no physical access to the product of interest
– Seller or buyer may decide not to commit the transaction
– Centralized: all information remains on eBay Servers
• Buyers and sellers evaluate each other after transactions
• The evaluation is not mandatory and will never be removed
• Each eBay member has a “reputation” (feedback score) that
is the summation of the numerical evaluations.
Andrew Diniz da Costa © LES/PUC-Rio
13
eBay model
Andrew Diniz da Costa © LES/PUC-Rio
14
eBay model
Andrew Diniz da Costa © LES/PUC-Rio
15
SPORAS & HISTOS
• Context: e-commerce, similar to eBay
• An individual may have a very high reputation in one
domain, while she has a low reputation in another.
• Two models are proposed:
– Sporas: works even with few evaluations (ratings)
– Histos: assumes abundance of evaluations
• Ratings given by users with a high reputation are weighted
more
• Reputation values are not allowed to increase at infinitum
Andrew Diniz da Costa © LES/PUC-Rio
16
SPORAS & HISTOS
• SPORAS
– Reputations are in [0, 3000]. Newcommers = 0. Ratings are in
[0.1, 1]
– Reputations never get below 0, even in the case of very bad
behaviours
– After each rating the reputation is updated
• HISTOS
– Aim: compute a global ‘personalized reputation’ value for each
member
Andrew Diniz da Costa © LES/PUC-Rio
17
Fire Model
• Three types of reputation
– Interaction trust
– Witness reputation
– Certified reputation
* Huynh, T. D., Jennings, N. R. and Shadbolt, N. (2004) FIRE: an integrated trust and reputation model for open multi-agent
systems. In: 16th European Conference on Artificial Intelligence, 2004, Valencia, Spain.
Andrew Diniz da Costa © LES/PUC-Rio
18
Fire Model
• Interaction trust
– resulting from past experiences from direct interactions
– Between [-1, +1]
– -1 means absolutely negative
– +1 means absolutely positive
– 0 means neutral or uncertain
Interaction Trust of the Agent B
(price, quality, etc)
Request
Agent A
Provide
Agent B
Andrew Diniz da Costa © LES/PUC-Rio
19
Fire Model
• Witness reputation
– reports of witness about an agent’s behaviour
Agent C
Agent D knows Agent B
Request witness
Agent A
Agent D
Agent B
Agent E
Andrew Diniz da Costa © LES/PUC-Rio
20
Fire Model
• Certified reputation
– references provided by other agents about its behaviour
Evaluation of D made
by the agent A
Evaluation of A made
by the agent D
Agent D
Evaluation of B made
by the agent A
0,5
Agent A
-0,5
Agent B
Evaluation of A made
by the agent B
Agent C
Andrew Diniz da Costa © LES/PUC-Rio
21
Governance Framework
- GUEDES, José ; SILVA, V. T. ; LUCENA, Carlos José Pereira de . A Reputation Model Based on Testimonies. In: Kolp, M,
Garcia, A, Ghoze, C, Bresciani, P, Henderson-Sellers, B, Mouratidis, M.. (Org.). Agent-Oriented Information Systems.:
Springer-Verlag, 2008, v. LNAI, p. 37-52.
- DURAN, Feranda ; SILVA, V. T. ; LUCENA, Carlos José Pereira de . Using Testimonies to Enforce the behavior of Agents.
In: Sichman, J., Noriega, P., Padget, J. and Ossowski, S.. (Org.). Coordination, Organizations, Institutions and Norms in
Agent Systems III. : Springer-Verlag, 2008, v. LNAI, p. 218-231.
Andrew Diniz da Costa © LES/PUC-Rio
22
Governance Framework – Reputation System
• Three different kinds of reputations were defined:
– role reputation, norm reputation and global reputation.
• Role reputations only consider norms that were violated
while playing a specified role or lies that were told while
playing the role.
• Norm reputations focus on the violation of a norm and on
the lies told while considering a norm.
• The global reputation of an agent considers all violated
norms and all told lies.
Andrew Diniz da Costa © LES/PUC-Rio
23
Presentation Outline
• Part 1: Introduction
– Motivation
– Some definitions
• Part 2: Computational trust and reputation models
– eBay/OnSale
– SPORAS & HISTOS
– Fire Model
– Governance Framework
• Part 3: ART-Testbed
– Overview
Andrew Diniz da Costa © LES/PUC-Rio
24
Domain
Andrew Diniz da Costa © LES/PUC-Rio
25
Reputation Transaction Protocol
Andrew Diniz da Costa © LES/PUC-Rio
26
Opinion Transaction Protocol
Andrew Diniz da Costa © LES/PUC-Rio
27
Simulator
Andrew Diniz da Costa © LES/PUC-Rio
28
Competition
• 17 agents (1 didn’t execute) of 13 different institutions
• Two phases
– Preliminary
– Final
• Preliminary phase (May 10-11)
– 8 agents of the different institutions
– 15 agents offered by competition (5 “bad”, 5 “neutral”, 5 “bad”
dummies )
– 100 rounds
• Final phase (May 16-17)
– 5 best agents of the preliminary phase
– 15 agents offered by competition (5 “bad”, 5 “neutral”, 5 “bad”
dummies )
– 200 rounds
Andrew Diniz da Costa © LES/PUC-Rio
29
Preliminary Phase
Andrew Diniz da Costa © LES/PUC-Rio
30
Final Phase
1) Electronics & Computer Science, University of Southampton
2) Department of Math & Computer Science, The University of Tulsa
3) Department of Computer Engineering, Bogazici University
4) Agents Research Lab, University of Girona
5) Pontifícia Universidade Católica do Rio de Janeiro
Andrew Diniz da Costa © LES/PUC-Rio
31
Conclusion
• ART-Testbed is being useful, however:
– What is reputation?
– Unreal Domain
• Researches have worked in domains of the industry to apply
trust and reputation.
• Area is growing
• Famous researches are working in this area.
Andrew Diniz da Costa © LES/PUC-Rio
32
References
•
•
•
•
•
•
•
•
•
•
[AbdulRahman, 97] A. Abdul-Rahman. The PGP trust model. EDI-Forum: the Journal of Electronic
Commerce, 10(3):27–31, 1997.
[Barber, 83] B. Barber, The Logic and Limits of Trust, The meanings of trust: Technical competence
and fiduciary responsibility, Rutgers University Press, Rutgers, NJ, United States of America, 1983,
p. 7-25.
[Carbo et al., 03] J. Carbo and J. M. Molina and J. {Dávila Muro, Trust Management Through Fuzzy
Reputation, International Journal of Cooperative Information Systems, 2003, vol. 12:1, p. 135-155.
[Casare & Sichman, 05] S. J. Casare and J. S. Sichman, Towards a functional ontology of
reputation, Proceedings of AAMAS’05, 2005.
[Castelfranchi, 00] C. Castelfranchi, Engineering Social Order, Proceedings of ESAW’00, 2000.
[Castelfranchi & Falcone, 98] C. Castelfranchi and R. Falcone, Principles of trust for MAS:
Cognitive anatomy, social importance and quantification. Proc of ICMAS’98, pages 72-79, 1998.
[Conte & Paolucci, 02] R. Conte and M. Paolucci, Reputation in Artificial Societies. Social Beliefs
for Social Order, Kluwer Academic Publishers, G. Weiss (eds), Dordrecht, The Netherlands, 2002.
[Dellarocas, 00] C. Dellarocas, Immunizing online reputation reporting systems against unfair
ratings and discriminatory behavior, p. 150-157, Proceedings of the ACM Conference on "Electronic
Commerce“ (EC'00), October, ACM Press, New York, NY, United States of America, 2000.
[Dellarocas, 01] C. Dellarocas, Analyzing the economic efficiency of {eBay-like} online reputation
reporting mechanisms, p. 171-179, Proceedings of the ACM Conference on "Electronic Commerce"
(EC'01), October, ACM Press, New York, NY, United States of America, 2001.
[Demolombe & Lorini, 08] R. Demolombe and E. Lorini, Trust and norms in the context of
computer security: a logical formalization. Proc of DEON’08, LNAI, 1998.
Andrew Diniz da Costa © LES/PUC-Rio
33
References
•
•
•
•
•
•
•
•
•
[Fullam et al, 05] K. Fullam, T. Klos, G. Muller, J. Sabater-Mir, A. Schlosser, Z. Topol, S. Barber, J.
Rosenschein, L. Vercouter and M. Voss, A Specification of the Agent Reputation and Trust (ART)
Testbed: Experimentation and Competition for Trust in Agent Societies, Proceedings of AAMAS’05,
2005.
[Herzig et al, 08] A. Herzig, E. Lorini, J. F. Hubner, J. Ben-Naim, C. Castelfranchi, R. Demolombe,
D. Longin and L. Vercouyter. Prolegomena for a logic of trust and reputation, submitted to Normas
08.
[Luhmann, 79] N. Luhmann, Trust and Power, John Wiley \& Sons, 1979. [McKnight & Chervany,
02] D. H. McKnight and N. L. Chervany, What trust means in e-commerce customer relationship: an
interdisciplinary conceptual typology, International Journal of Electronic Commerce, 2002.
[Mui et al., 02] L. Mui and M. Mohtashemi and A. Halberstadt, Notions of Reputation in Multi-agent
Systems: A Review, Proceedings of Autonomous Agents and Multi-Agent Systems (AAMAS'02), p.
280-287, 2002, C. Castelfranchi and W.L. Johnson (eds), Bologna, Italy, July, ACM Press, New York,
NY, United States of America.
[Muller & Vercouter, 05] G. Muller and L. Vercouter, Decentralized Monitoring of Agent
Communication with a Reputation Model, Trusting Agents for trusting Electronic Societies, LNCS
3577, 2005.
[Pearl, 88] Pearl, J. Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference,
Morgan Kaufmann, San Francisco, 1988.
[Rehák et al., 05] M. Rehák and M. Pěchouček and P. Benda and L. Foltn, Trust in Coalition
Environment: Fuzzy Number Approach, Proceedings of the Workshop on "Trust in Agent Societies" at
Autonomous Agents and Multi-Agent Systems (AAMAS'05), p. 132-144, 2005, C. Castelfranchi and
S. Barber and J. Sabater and M. P. Singh (eds) Utrecht, The Netherlands, July.
[Sabater, 04] Evaluating the ReGreT system Applied Artificial Intelligence, 18 (9-10) :797-813.
[Sabater & Sierra, 05] Review on computational trust and reputation models Artificial Intelligence
Review ,24 (1) :33-60.
Andrew Diniz da Costa © LES/PUC-Rio
34
References
•
•
•
•
•
•
[Sabater-Mir & Paolucci, 06] Repage: REPutation and imAGE among limited
autonomous partners, JASSS - Journal of Artificial Societies and Social Simulation ,9
(2), 2006.
[Schillo & Funk, 99] M. Schillo and P. Funk, Learning from and about other agents in
terms of social metaphors, Agents Learning About From and With Other Agents, 1999.
[Sen & Sajja, 02] S. Sen and N. Sajja, Robustness of reputation-based trust:
Boolean case, Proceedings of Autonomous Agents and Multi-Agent Systems
(AAMAS'02), p. 288-293, 2002, Bologna, Italy, M. Gini and T. Ishida and C.
Castelfranchi and W. L. Johnson (eds), ACM Press, New York, NY, United States of
America, vol.1.
[Shapiro, 87] S. P. Shapiro, The social control of impersonal trust, American Journal
of Sociology, 1987, vol. 93, p. 623-658.
[Steiner, 03] D. Steiner, Survey: How do Users Feel About eBay's Feedback System?
January, 2003, http://www.auctionbytes.com/cab/abu/y203/m01/abu0087/s02 .
[Zacharia et al., 99] G. Zacharia and A. Moukas and P. Maes, Collaborative
Reputation Mechanisms in Electronic Marketplaces, Proceedings of the Hawaii
International Conference on System Sciences (HICSS-32), vol. 08, 1999, p. 8026,
IEEE Computer Society, Washington, DC, United States of America.
Andrew Diniz da Costa © LES/PUC-Rio
35
Perguntas...
Andrew Diniz da Costa
[email protected]
Andrew Diniz da Costa © LES/PUC-Rio
36
Download

Trust_Reputation2010.. - (LES) da PUC-Rio