EvoSec 2025W Lecture 16
Readings
- Li, "Securing Email Archives through User Modeling." (ACSAC 2005)
- Li, "Fine-grained Access Control using Email Social Networks." (CATX 2013)
Discussion Questions
Feel free to only address a subset or none of the following questions in your discussion!
- What does it take to define "normal"? In what contexts is it easier to define normal, and where is it harder?
- To what extent does improved technology make it easier to distinguish between normal and abnormal behavior in an adversarial context?
- When are false alarms okay, and when are they bad? (How often do you get alerts today from security systems and how often are these irrelevant?)
- In general, is it better to look at data or metadata when doing anomaly detection?
- How does the metadata for modern communication platforms differ from email? How is it similar?
Notes
Lecture 16
----------
G1
- defining normal: long term, consistent data collected
- hard to define normal if there isn't enough data (not enough interaction)
or consistency
- don't need too much consistency, just need lots of data so the patterns
can be extracted
- false alarms
- notifications from google regarding logins is mostly false alarms but is still useful for maintaining security
- but too many notifications for regular activity will lead to users to ignore, so frequency matters
- severity also matters, can stress out users for no reason
G2
- "window length" idea, how to apply generally?
- harder to define normal with smaller & smaller window lengths
- larger behavior space, more possible actions makes it harder to define normal
- false alarms: intensity of alarms matters, how easy to ignore/how concerning
- data vs metadata: generally metadata is the way to go
- context matters, hard to get context from data
(but not always, e.g., topics)
G3
- LLM/AI could help with looking at data for determining normal/classification
- is it good to get all of this data? could be used for impersonation
- false alarms
- geographic change-based alerts can be reasonable, for example
- but could dissuade users from trying new things
- 2FA on Carleton email, how useful when done on same device?
- not good when it is hard to access important information quickly, interferes with normal tasks
- modern platforms track more: typing, geographic info
- more invasive
- gets confused - IP address is in Toronto but still in Ottawa
- assigning tasks/roles - using just email can be too limited
- but modern platforms are controlled by large companies, so can see
info across apps
G4
- does better tech make defining normal easier?
- not really!
- newer tech, newer kinds of abnormalities
- still an open problem in machine learning 20 years later!
- still not getting great accuracy
- machine learning black boxes don't help so much for anomaly detection
- people need to go deeper to derive mathematical relationships
- have to look at the metadata to determine attacks, don't have ground truth
(don't know what is really an attack) most of the time
- if attacker has information (e.g., emails), can mask their attacks, hide from detection
Easy to do anomaly detection wrong
- focus on modeling everything, rather than what you must model
- no clear idea of what "normal" will be
machine learning is best used first as a tool for data exploration
- can use in production, but ONLY after you really understand what it does
- machine learning isn't always the best at identifying features!
- because it lacks context
to do security well, you need "normal" to be very consistent
- which means humans should be able to do the classification relatively easily
So the art of this is to figure out what will be consistent
- use domain knowledge & machine learning exploration of data
email archive detection
- attacker evasion is either noticable to automated system or user
Work backwards from attacks!
- why are they weird?