Employees adhere more to unethical instructions from human than AI supervisors: complementing experimental evidence with machine learning
The role of artificial intelligence (AI) in organizations has fundamentally changed from performing routine tasks to supervising human employees. While prior studies focused on normative perceptions of such AI supervisors, employees' behavioral reactions towards them remained largely unexplored...
Authors: | ; ; |
---|---|
Format: | Electronic Article |
Language: | English |
Check availability: | HBZ Gateway |
Journals Online & Print: | |
Fernleihe: | Fernleihe für die Fachinformationsdienste |
Published: |
Springer
2024
|
In: |
Journal of business ethics
Year: 2024, Volume: 189, Issue: 3, Pages: 625-646 |
Further subjects: | B
Perceived mind
B Artificial Intelligence B Aufsatz in Zeitschrift B AI leadership B Unethical leadership |
Online Access: |
Volltext (kostenfrei) Volltext (kostenfrei) |
Summary: | The role of artificial intelligence (AI) in organizations has fundamentally changed from performing routine tasks to supervising human employees. While prior studies focused on normative perceptions of such AI supervisors, employees' behavioral reactions towards them remained largely unexplored. We draw from theories on AI aversion and appreciation to tackle the ambiguity within this field and investigate if and why employees might adhere to unethical instructions either from a human or an AI supervisor. In addition, we identify employee characteristics affecting this relationship. To inform this debate, we conducted four experiments (total N = 1701) and used two state-of-the-art machine learning algorithms (causal forest and transformers). We consistently find that employees adhere less to unethical instructions from an AI than a human supervisor. Further, individual characteristics such as the tendency to comply without dissent or age constitute important boundary conditions. In addition, Study 1 identified that the perceived mind of the supervisors serves as an explanatory mechanism. We generate further insights on this mediator via experimental manipulations in two pre-registered studies by manipulating mind between two AI (Study 2) and two human supervisors (Study 3). In (pre-registered) Study 4, we replicate the resistance to unethical instructions from AI supervisors in an incentivized experimental setting. Our research generates insights into the "black box" of human behavior toward AI supervisors, particularly in the moral domain, and showcases how organizational researchers can use machine learning methods as powerful tools to complement experimental research for the generation of more fine-grained insights. |
---|---|
ISSN: | 1573-0697 |
Contains: | Enthalten in: Journal of business ethics
|
Persistent identifiers: | DOI: 10.1007/s10551-023-05393-1 |