Many students expressed concerns about displaying personal information such as names, jobs, or other contextual data without the person’s knowledge or consent.
The exposure of information in public spaces—even if it is publicly available—was seen as problematic when used in real-time recognition without explicit permission.
📊 Data Security and Misuse
Several students mentioned the risk of data misuse, including scenarios where sensitive or outdated data might be accessed or misinterpreted.
Concerns were raised about the lack of control over data once it is accessed via MR glasses, including who stores it and how it might be shared.
📡 Consent and Transparency
A recurring theme was the need for informed consent—people should be aware that their data is being accessed and have control over what is shared.
Students suggested that transparency of sources (e.g., which database the information comes from) is crucial to build trust and ensure ethical usage.
🧠 Bias and Discrimination
There were mentions of the potential for algorithmic bias, especially if data is selectively filtered or interpreted through flawed systems.
Students worried that certain demographics could be unfairly targeted or judged based on the type or presentation of information shown.
⚠️ Accuracy and Reliability
Students warned about inaccurate or outdated data being displayed, which could misrepresent the person and lead to harmful consequences.
Emphasis was placed on the need for reliable, verified sources, especially if information could influence perceptions in social or professional contexts.