Newswise — Researchers at the University of Notre Dame are developing artificial intelligence tools that help consumers understand how they are being exploited as they navigate online platforms. The goal is to boost the digital literacy of end users so they can better control how they interact with these websites.
In a recent study, participants were invited to experiment with online privacy settings without consequence. To test how different data privacy settings work, the researchers created a Chrome browser plug-in called Privacy Sandbox that replaced participant data with personas generated by GPT-4, a large language model from OpenAI.
With Privacy Sandbox, participants could interact with different websites, such as social media platforms or news outlets. As they navigated to various sites, the browser plug-in applied AI-generated data, making it more obvious for participants to see how they were targeted based on their supposed age, race, location, income, household size and more.
“From a user perspective, allowing the platform’s access to private data may be appealing because you could get better content out of it, but once you turn it on, you cannot get that data back. Once you do, the site already knows where you live,” said Toby Li, assistant professor of computer science and engineering and a faculty affiliate at the Lucy Family Institute for Data & Society at Notre Dame, who led the research. “This is something that we wanted participants to understand, figure out whether the setting is worth it in a risk-free environment, and allow them to make informed decisions.”
Another study looked at dark patterns — or the design features on digital platforms that subtly nudge users to perform specific actions — and how they are used on websites to manipulate customers. For the study, Li and his team looked at how dark patterns are applied by interface designers to encourage people to consume more content or make impulsive purchasing decisions.
The researchers developed a Chrome browser plug-in dubbed Dark Pita to identify dark patterns on five popular online platforms: Amazon, YouTube, Netflix, Facebook and X.
Using machine learning, the plug-in would first notify study participants that a dark pattern was detected. It would then identify the threat susceptibility of the dark pattern and explain the impact of the dark pattern — financial loss, invasion of privacy or cognitive burden. Dark Pita would then give participants the option to “take action” by modifying the website code through an easy-to-use interface to change the deceptive design features of the site and explain the effect of the modification.
The researchers plan to eventually make both browser plug-ins, Privacy Sandbox and Dark Pita, available to the public. Li believes these tools are great examples of how the use of AI can be democratized for regular users to benefit society.
“Companies will increasingly use AI to their advantage, which will continue to widen the power gap between them and users. So with our research, we are exploring how we can give back power to the public by allowing them to use AI tools in their best interest against the existing oppressive algorithms. This ‘fight fire with fire’ approach should level the playing field a little bit,” Li said.
“An empathy-based sandbox approach to bridge the privacy gap among attitudes, goals, knowledge, and behaviors” was presented this month at the 2024 Association of Computing Machinery CHI Conference. Led by Li, fellow study co-authors include Chaoran Chen and Yanfang (Fanny) Ye from Notre Dame, Weijun Li from Zhejiang University, Wenxin Song at the Chinese University of Hong Kong and Yaxing Yao at Virginia Tech.
The study “From awareness to action: Exploring end-user empowerment interventions for dark patterns in UX,” led by Li, has been published in the Proceedings of the ACM on Human-Computer Interaction (CSCW 2024). Co-authors include Yuwen Lu from Notre Dame, Chao Zhang from Cornell University, Yuewen Yang from Cornell Tech and Yao from Virginia Tech.
This research was funded by a Google Research Scholar Award, a Google PSS Privacy Research Award and the National Science Foundation.
Contact: Brandi Wampler, associate director of media relations, 574-631-2632, [email protected]