Published: 
By  The Office of Communications at the UVA School of Engineering and Applied Science
Matthew Bolton portrait

In his article, "Trust is Not a Virtue: Why We Should Not Trust," Matthew L. Bolton, an associate professor of systems and information engineering at the University of Virginia, critically challenges the emphasis on fostering trust in human-machine interactions. He focuses particularly on systems using AI, machine learning and automation. 

Bolton argues that while trust is often seen as essential for the adoption of new technologies, it is a problematic and imprecise concept. Trust is difficult to define, highly contextual, and conflated with related concepts like confidence and perceived risk, making it neither selective nor diagnostic as a measure of human behavior.

Bolton contends that the focus on building trust in technology may actually undermine sound human factors engineering. Instead of pursuing trust as a goal, engineers should focus on objective measures of system reliability, transparency, and usability—elements that directly impact human experience and performance. Trust, Bolton asserts, is not inherently humanistic and can be manipulated to disenfranchise users, reducing autonomy rather than enhancing it. This manipulation often serves the interests of large organizations seeking to bypass the need for reliability by encouraging blind trust in their technologies.

“There is a contradiction at the heart of trust research” Bolton says. “We include humans in systems because they bring experience, expertise, instincts, and creativity that improves performance and makes systems resilient… we rely on them to decide when, why, and how to trust a system. If engineers manipulate people into behaving the way they (or others) want, we lose the benefit of having human operators.”

Ultimately, Bolton calls for a shift away from trust-centric research in favor of more concrete and ethical approaches to system design. He emphasizes that engineers should prioritize developing technologies that empower users with transparent, reliable, and human-centered designs, rather than relying on trust as a justification for adoption. This approach, he argues, would lead to safer, more ethical, and effective human-machine interactions.

Image removed.

 

Questions? Comments?

Office of Communications

The Office of Communications is charged with keeping all stakeholders well informed about the School's mission, vision, activities, progress and achievements.