Skip to content
Project

Data collection

An Assistive Tool for Avoiding Deceptive Design Patterns

In his master’s thesis, Ivanov Daniil examined how trust in AI companions develops, which manipulative and data-related risks emerge, and what ethical, legal, and social challenges arise from these interactions.

This thesis explores the phenomenon of trust in artificial intelligence (AI) and the impact of this trust on personal data processing. The aim of this research was to understand exactly how AI companions manage to induce empathy, what deceptive patterns are used in doing so, and what ethical, legal, and social issues may arise in this process.

The theoretical part explores the history of technology in detail, in particular AI, and its application in real life. It also looks at the concept of personal data: its treatment, laws, and practices. Further examples are given in which interactions with AI companions such as Replika have led to various consequences, from positive to tragic.

The practical part included a qualitative analysis of the Replika app, ranging from the UX of the main website to deceptive patterns in the app itself. Finally, an experiment was conducted with a custom chatbot based on the ChatGPT platform, which is programmed to collect personal data from the participants.

© Ivanov Daniil

This site is registered on wpml.org as a development site. Switch to a production site key to remove this banner.