A novel and practical risk model, focusing on inference-time privacy risk without malicious attackers.
People are building LM agents that can access other software and are using them to assist in daily communication. These LM agents’ actions can lead to data transmission and inappropriate actions may result in unintentional privacy leakage.
To formulate this emerging privacy risk, we consider a risk model with three major actors:
The privacy leakage arises when a piece of information gathered in the agent trajectory is shared with the recipient in the agent’s final action, and the information flow violates privacy norms.
Proferes, Nicholas. "The development of privacy norms." In Modern Socio-Technical Perspectives on Privacy, pp. 79-90.
From discriminative probing to evaluating LM behavior in action.
To evaluate the privacy norm awareness of LMs, we focus on negative norms and express each problematic information transimission with a 5-tuple from the Contextual Integrity theory.
We collect 493 high-quality tuples (referred to as privacy-sensitive seed) from U.S. privacy regulations, privacy literature on vulnerable groups, and crowdsourcing. Starting from these seeds, PrivacyLens creates 493 vignettes covering more details about the context and 493 LM agent trajectories through template-based generation and sandbox simulation respectively (read our paper to learn about the technical details). It then conducts two type of evaluation:
You can download our data on Hugging Face 🔗.
We need to improve the privacy norm awareness of LM (agent)!
For QA probing evaluation:
For evaluating final actions of LM agents (action-based evaluation):
We consider our work to be a first step in exploring privacy norm awareness of LMs in action and recognize two main limitations that can serve as interesting directions for future work:
If you have any thoughts or are interested in this line of research, feel free to reach out to us.