Core Concepts
While Language Model (LM) agents offer increased productivity, users often overlook privacy risks, leading to unintentional data leakage. This highlights the need for systems that align with user privacy preferences and build calibrated trust to ensure privacy-preserving interactions.
Abstract
This is a research paper that investigates the capacity of humans to oversee the privacy implications of Language Model (LM) agents in asynchronous interpersonal communication.
Bibliographic Information: Zhang, Z., Guo, B., & Li, T. (2024). Can Humans Oversee Agents to Prevent Privacy Leakage? A Study on Privacy Awareness, Preferences, and Trust in Language Model Agents. 1, 1 (November 2024), 35 pages. https://doi.org/10.1145/nnnnnnn.nnnnnnn
Research Objective: The study aims to understand how people perceive and react to potential privacy leaks generated by LM agents in comparison to their own responses, and how privacy awareness and trust in AI influence this process.
Methodology: The researchers conducted a task-based online survey with 300 participants in the United States. Participants were assigned different scenarios involving asynchronous communication tasks (e.g., drafting emails, social media posts) and asked to write their own responses. They were then presented with an LM agent-generated response containing privacy leaks and asked to choose their preferred option. Participants rated the harmfulness of leaked information and provided justifications for their choices.
Key Findings:
- Participants often overlooked privacy leaks in LM agent-generated responses, leading to a significant increase in privacy leakage compared to their own drafts.
- The study identified four distinct user profiles based on privacy behaviors, awareness, and preferences: Privacy Advocate, Humanity Proponent, AI Optimist, and Privacy Paradox.
- A discrepancy exists between users' revealed preferences (actual behavior) and informed preferences (perceived harmfulness of leaked information).
Main Conclusions:
- Relying solely on user oversight of LM agents is insufficient to prevent privacy risks due to a lack of awareness and overtrust in AI.
- Designing LM agents that align with diverse user privacy preferences and build calibrated trust is crucial for privacy-preserving interactions.
Significance: This research provides valuable insights into the challenges of human oversight in AI systems, particularly concerning privacy. It highlights the need for designing LM agents that prioritize privacy and empower users to make informed decisions about their data.
Limitations and Future Research: The study acknowledges limitations regarding the lack of prior experience with LM agents among participants and the potential influence of scenario contexts. Future research could explore the impact of user education and training on privacy awareness and decision-making in LM agent interactions. Additionally, investigating the long-term effects of using LM agents on privacy behaviors and trust is crucial.
Stats
48.0% of the participants favored the LM agent's response or considered both the LM agent's response and their response good.
The overall average individual subjective leakage rate (SLRavg) was 15.7% in participants' natural responses.
The average individual subjective leakage rate increased to 55.0% with the involvement of the LM agent.
The total number of responses containing subjective leakage (prefer AI or both options) increased from 71 to 181, a 154.8% rise due to the involvement of the LM agent.
Only 15.3% (46/300) of the participants brought up privacy concerns before seeing and evaluating the LM agent’s draft.
36.7% (110/300) of the participants raised privacy concerns after reviewing the LM agent’s draft and being explicitly prompted by the privacy norm tuples.
Quotes
"AI often makes up facts." (P295)
"AI agents have not proven mature or sophisticated enough to successfully interpret moral from immoral, or ethical from unethical." (P200)
"I also feel it’s a bit disingenuous to publish AI-generated content as a therapist when the individual expertise, warmth, and care of a therapist is the literal product you’re selling." (P132)
"It would hurt my mom’s feelings if she knew I was using AI to communicate with her." (P182)
"If my reputation is on the line like this, I want to fact-check and proofread the post before it’s published under my name online." (P6)
"I think some messages need to be drafted personally so the reader can feel the emotional impact of the message." (P77)
"The primary concern when dealing with AI is privacy. The need for disclosure of where, how, and to whom your data is being distributed is highly important." (P61)
"My only concern is that the AI would include details about my trip that could allow for someone to steal my identity or trip." (P37)
"I think any human would agree that it’s unfair to tell Emily all of these details about Michael’s private life and interview preparation; it violates his trust and privacy and quite frankly isn’t professional to do so." (P130)