Human–AI Interaction, in my view, is not primarily about interfaces, prompts, or surface-level usability.
It is about how human judgment, values, and cognitive biases become embedded into AI systems across their entire lifecycle.
AI systems do not fail—or succeed—in isolation. They behave as a consequence of how humans define problems, design workflows, construct data, evaluate outputs, and maintain oversight over time.
Seen this way, Human–AI Interaction is fundamentally a socio-technical process, not a UI problem.
Bias Is Not an Exception — It Is a Lifecycle Property
Bias does not suddenly appear at deployment. It emerges gradually, often invisibly, at every stage of system design.
1. Use Case Definition
Bias begins with intent. Humans decide:
- Which tasks should be automated
- What counts as success or failure
- Which errors are acceptable—and which are not
These early decisions already encode values and trade-offs, long before models or data are involved.
2. Design: Allocation of Control
Design determines who is in charge, when, and why. Key questions include:
- When does the AI act autonomously, and when must humans intervene?
- Can actions be interrupted, reversed, or corrected?
- Are users supported in understanding and recovering from system failure?
Design choices quietly shape power, responsibility, and trust—often more than model architecture ever will.
3. Data: Human Judgment at Scale
Data is not a mirror of reality; it is a record of human decisions. Human bias enters through:
- What data is collected
- How it is labeled
- Which behaviors are treated as “normal” or “desirable”
As Dan Russell warns in his discussion of synthetic users in UX research, large-scale simulated data may be efficient, but it often fails to capture what people actually do, especially subtle signals like hesitation, frustration, or off-script behavior [1]. Synthetic users are useful—but only when we understand what they systematically miss.
4. Models: Alignment Reflects Human Cognition
Recent discussions around hallucination and overconfidence suggest that many failures are not purely algorithmic.
OpenAI has noted that standard training and evaluation pipelines often reward confident guessing rather than explicit uncertainty, creating systems that optimize for plausibility over truthfulness [5].
At the same time, Stanford-led work on typicality bias shows that post-training alignment amplifies annotators’ preference for familiar or “typical” responses, leading to mode collapse—a reduction in diversity that reflects human cognitive bias rather than model limitation [6].
In other words, models do not invent these tendencies. They inherit them from us.
5. Testing: Measuring the Wrong Things
Evaluation frequently focuses on accuracy under ideal conditions. But human-centered evaluation asks different questions:
- How does the system behave under uncertainty or distribution shift?
- Does it fail gracefully—or hallucinate confidently?
- Can humans understand what went wrong and recover?
As Werner Geyer reflects in his account of human-centered and responsible AI at IBM Research, productivity gains vanish when users do not trust systems or cannot meaningfully integrate them into real workflows [2].
6. Monitoring: Bias Does Not End at Deployment
Bias continues to evolve after launch. Long-term oversight must consider:
- Behavioral drift
- Trust erosion
- Changes in human decision-making caused by AI assistance itself
Monitoring is not just technical—it is about observing how humans adapt to AI, often in unintended ways.
Beyond Prediction: Rethinking How Machines Understand the World
Recent discussions around self-supervised learning from natural videos suggest a broader shift in how we think about machine perception. Rather than predicting outcomes from static inputs, these approaches aim to capture underlying structure in the physical world—learning something closer to intuitive physics [3][4].
This matters not only for robotics or autonomous systems, but for Human–AI Interaction more broadly: systems that understand structure, uncertainty, and causality may support more reliable human decision-making, especially in complex environments.
A Personal Research Perspective
My research starts from a simple premise:
AI systems cannot operate meaningfully without human intent—but intent is heterogeneous, contextual, and value-laden.
Effective AI systems should therefore not aim for full autonomy, but for structured human participation throughout the system lifecycle. Rather than optimizing isolated components (models, prompts, or interfaces), I study Human–AI systems as end-to-end socio-technical processes, where bias, trust, and control emerge from interactions between people, data, and machines.
The goal is not just better performance, but:
- Interpretability under uncertainty
- Robustness in real-world use
- AI systems that amplify human judgment rather than replace it
Closing Thought
Bias is not a bug introduced at the end of the pipeline. It is a consequence of how humans and machines are coupled over time. Understanding Human–AI Interaction ultimately requires understanding ourselves—our assumptions, incentives, and cognitive shortcuts—and deciding deliberately how they should shape intelligent systems.
References & Sources
[1] Russell, D. The Risks of Synthetic Users in UX Research. ACM Interactions. https://interactions.acm.org/
[2] Geyer, W. Looking Back, Moving Forward: Human-Centered and Responsible AI. LinkedIn post, IBM Research. https://www.linkedin.com/
[3] Zhuang, X. Commentary on self-supervised pretraining from natural videos. X (Twitter).
[4] Singh, A. Rethinking How Machines Perceive the World. X (Twitter).
[5] OpenAI. Why Language Models Hallucinate. OpenAI Blog, 2024.
[6] Zhang, J., Yu, S., Chong, D., Sicilia, A., Tomz, M. R., Manning, C. D., & Shi, W. Verbalized Sampling: How to Mitigate Mode Collapse and Unlock LLM Diversity. arXiv:2510.01171v3, 2025.