EvoSec 2025W Lecture 3: Difference between revisions
Created page with "<pre> Lecture 3 --------- Perspectives on Trust G1 - waking up - do you trust that nothing bad will happen, or you just get up because you have to? - we can decouple, but can machines decouple trust from action? - continuous vs discrete trust - "levels of trust" - how does that affect actions G2 - game theory - prisoner's dilemma, agents are adversaries? where is the sociality of trust - not a full view - trust as black and white vs probability G3 -..." |
(No difference)
|
Latest revision as of 16:48, 15 January 2025
Lecture 3 --------- Perspectives on Trust G1 - waking up - do you trust that nothing bad will happen, or you just get up because you have to? - we can decouple, but can machines decouple trust from action? - continuous vs discrete trust - "levels of trust" - how does that affect actions G2 - game theory - prisoner's dilemma, agents are adversaries? where is the sociality of trust - not a full view - trust as black and white vs probability G3 - trust from creators to machines we create - humans have a sense of self preservation, that's where trust comes from - computers don't have that do they? G4 - humans develop trust over time - computers make trust decisions instantly - humans & computers value different things for trust - IP address vs image/voice in video call Why did I assign these readings? - see the variety - note the lack of coherent purpose game theory, prisoner's dilemma - cooperation is a bad idea - unless it is an iterated game - classic strategy, tit-for-tat - best strategy is tit-for-tat with revenge (for preserving cooperation) - but what if you already trust each other, why would you defect? Analysis vs synthesis - computer security is on the synthesis side Why do computer systems make trust decisions - so fast - without reference to past experience - too much work/effort/computation/storage? - user makes the trust decision? Traditionally, computers haven't been empowered to make trust decisions - instead, they enforce trust relationships decided by people - i.e. people create policy, computers enforce policy - in those policies, the trust associated to entities is NOT determined by the computer Why? - computers aren't autonomous Only autonomous systems can really be trusted - because they are empowered to say no What secrets will a computer never reveal? - secret keys is TPMs and similar devices