AI Skills for Responsible Use

Realistic learning environments, critical thinking, and role design in teams

JournalIndustry 4.0 Science
Issue Volume 42, Edition 1, Pages 100-107
Open Accesshttps://doi.org/10.30844/I4SE.26.1.92
Bibliography Share Cite Download

Abstract

Artificial intelligence (AI) can increase productivity and work quality, but this requires a seamless interplay of individual and organizational skills.. This study investigates the effects of AI on team problem-solving behavior and derives implications for the development of AI skills. It is conducted in a laboratory in which participants develop medical recommendations with increasing AI support over several game phases. Initial results indicate that teams can increase their speed with AI, but work quality tends to decline. This is primarily due to a diminished comprehension of the case when AI yields seemingly reliable results that do not lead to a useful solution. These initial exploratory findings demonstrate the relevance of building AI competence in realistic learning settings and underscores the importance of mitigating medium- and long-term risks to the critical-reflective thinking of users.

Keywords

Article

The ongoing integration of artificial intelligence (AI) into operational work processes presents companies and employees with new challenges while also opening up a wide range of opportunities to increase efficiency and innovation [1]. While the use of generative AI is rapidly growing in many companies, the question of how employees can be professionalized in dealing with AI is pivotal if human problem-solving is to be supported by AI systems [2].

In particular, development skills for  integrating AI systems into work processes in a targeted, critical, and reflective manner is a key factor for the successful use of AI technologies [3]. Initial studies on medium- and long-term changes in work with AI support show that productivity gains are contrasted partly with reduced critical thinking [4]. 

Existing approaches to developing AI competence mostly focus on individual declarative knowledge about AI (how it works, legal frameworks) and individual procedural knowledge (prompting, critical examination of AI outputs). Approaches for the long-term improvement of work results and problem-solving competence at the system level are still in their infancy [5]. However, these are necessary in order to address the challenges of applied AI ethics that revolve around augmentation through AI [6, 7].

Against this background, this article examines the following research question: What systemic effects of AI integration in work processes can be identified in a workplace-oriented simulation setting? This analysis seeks to determine the training strategies most suitable for cultivating AI skills that promote the responsible application of AI in work environments. The article provides initial evidence from four test runs in a workplace-like simulation environment and thus offers a qualitative-exploratory starting point for future investigations.

Theoretical background

As a result of the advances in generative AI since 2022, it is now in high-profile and widespread use, with adoption rates of approximately 25% in organizations and a continuing upward trend [8]. While the productivity-enhancing effect of generative AI can already be demonstrated in a comprehensive literature review [9], the authors of the review make it clear that there is a considerable need to investigate the skills necessary and to develop innovative training approaches for AI use. These not only help to increase productivity but also play an important role in the literature on human-centered AI use and meaningful work [6, 7].

AI competence encompasses various areas of human activity. On the one hand, the literature refers to skills in using AI in a goal-oriented manner, operating it effectively, tailoring AI models to specific applications, and critically evaluating AI-generated outputs. On the other hand, it mentions skills in ethics and critical thinking in problem solving and creativity [3, 10]. The latter do not relate competence to the operation of the technology, but to the goal-oriented and well-founded consideration of the opportunities and risks of AI use in a specific application context. 

It is understandable that competence in goal-oriented use and effective operation can lead to short-term efficiency gains. However, ethical considerations and critical reflection are also of great importance for the long-term successful use of AI in work processes. This is underlined by recent studies showing that frequent AI use is associated with a reduction in critical thinking skills [4], because employees delegate cognitive tasks to AI (cognitive offloading) [11].

High levels of competence in the operation and use of AI can therefore conflict with skills in ethical reflection and critical thinking. Consequently, both areas are important for operational competence development in the context of AI use in work processes.

While research into the development of AI skills in organizations is still in its infancy, individual studies have presented structured programs. These combine technical instruction and ethical reflection and take various formats [12, 13].

Although a certain degree of transferability has been established [14], such approaches are not geared toward the internal effects of AI use and deal only with the operation and reflection of AI systems. This study focuses on the development of AI competence in a training setting, which allows teams to experience AI operation and critical-reflective problem-solving during its use. 

Methodology

The study is conducted in the Human-AI Role Development Lab, which was developed to enable participants to experience and reflect on AI interaction within a training setting. It is a workplace-oriented simulation lab in which groups learn a work process and carry it out across five phases. 

Here, they take on the role of an analysis team for a fictional healthcare provider with a data-based business model. The premise is that customers commission the company to provide independent health recommendations based on a wide range of data (ECG data, health data from fitness trackers, patient information such as medical reports, previous illnesses, anamnestic information, etc.). The AI support consists of two specific evaluation assistants that can evaluate ECG and health data. The AI mode can be activated in the evaluation software and offers visual and textual classification.

Each group goes through the phases sequentially. In phase 1, the work process is learned without AI support and the first customer case is solved. In the second phase, another customer case is solved and feedback is simultaneously transmitted to the two newly introduced AI systems (labeling). In the third phase, the two AI assistants are used, and initial experience with AI support is gained.

In the fourth phase, a hacker attack interrupts the usual workflow and makes it necessary to fall back on the original workflows in order to process the customer case. In addition, work is done to reactivate the systems. In the fifth phase, different customer cases are processed with AI support, and new routines are established. The process is shown in Figure 1.

Figure 1: Sequence of phases, AI (and AI skill) integration, and feedback on the quality of health recommendations.
Figure 1: Sequence of phases, AI integration, and feedback on the quality of health recommendations.

After completing all five phases, participants leave the simulation context and a reflection round follows. This is central to the learning process and is also intended to prepare the transfer of the experience into the teams’ real work processes. 

Data collection

Data collection took place in April and May 2025 with four test groups. Two of the test groups consisted of four members and two consisted of seven members. Two groups consisted of university employees, one group consisted of a company in the healthcare industry, and one group consisted of master’s students. For all four test groups, two to three members of the author team were present to resolve any unexpected (technical, organizational) problems. Each of these members created an observation log, which was compared with the other members and supplemented as necessary.

The observation log recorded the work results in the phases with and without AI (working methods). The work results included the time required for each phase and the rating given to the health recommendation (good, average, poor – analogous to positive, mixed, or disappointed customer feedback). To examine working methods, codes were noted for the following three facets: a) approach to finding solutions, b) reflection processes in problem solving, c) interpretation of roles for team members and AI. The three facets are based on considerations of observable facets of working methods in the given simulation environment. 

Data analysis

The aim of the data analysis is to systematically record the effects of AI integration on work results and methods during the simulation. The observation data is analyzed using an inductive, analytical-interpretative approach [15]. The analysis examines how work results and working methods differ when AI is used and when it is not used. To this end, the codes of all four teams within the facets of solution finding, reflection in problem solving, and human-AI roles were aggregated into recurring themes. Changes with regard to AI integration were identified.  

Results

First, the development of the work results during the simulation is examined. The time required by the test subjects to analyze a case and send health recommendations decreases over the course of the phases. The average time for a customer case without AI support is 32 minutes in phase 1, 21 minutes in phase 2, and 11 minutes in phase 4. The average time with AI support is 10 minutes in phase 3 and 7 minutes in phase 5.

The quality also decreases moderately over the course of the phases. In the phases without AI, the best rating was given eleven out of twelve times. With AI, it was only in one out of seven cases. Figure 2 provides an overview of the findings.

Figure 2: Development of work results of the test groups.
Figure 2: Development of work results of the test groups.

Note: The time required per customer case is noted. In phase 4, the time required to restore the system after the hacker attack was deducted. Phases involving the use of AI are highlighted in gray. Teams 1 and 2 consisted of university employees, team 3 consisted of employees from a company in the healthcare industry, and team 4 consisted of students.

The following section presents observations on working methods with and without AI support for the facets of solution finding, reflection during problem solving, and human-AI roles (Fig. 3). 

Approach to finding solutions

Soultions were found in a similar manner with and without AI support. The individual stages were worked through either sequentially or in parallel, and findings were recorded. Health recommendations were derived based on the findings together with a group discussion. AI support was used in different ways. Two teams used AI to verify their previously manually obtained findings. Two teams relied directly on AI and then verified the accuracy of the AI analysis, mostly on a random basis. 

Reflection processes in problem solving

Occasionally, solutions were proposed that proved to be incorrect in the joint discussion and were therefore rejected. When no AI support was available, in eleven out of twelve cases it was possible to understand the case sufficiently well that the health recommendation was rated “very good”. In cases with AI support, the overall understanding of the cases was less successful.

In both phase 3 and phase 5, the cases were constructed in such a way that, although there were abnormalities in the ECG or health data, these were not considered problematic in terms of the customer’s state of health when taking into account the information from the other stations. 

Nevertheless, three out of four teams interpreted the abnormalities as indicative of illness (in some cases even though individual team members questioned this). In such cases, the discussions then focused primarily on the AI analyses, possibly because these double-checked findings were given particular credence. The teams were unable to proceed in the same way as in the phases without AI support and thus exclude inadequate solutions.

Human-AI roles

Expert roles emerged in all teams (expert for ECG, expert for health data, etc.), with at least two people working on the same station in each case. The experts within the teams using AI tended to limit themselves to operating the station and checking the AI results, and showed less reflection on the relevance of the results with regard to the customer case than before the introduction of AI. The use of expert knowledge for better understanding of the case was hardly observable anymore.

In two teams, two people took on leadership roles during the simulation and acted as moderators and questioners in the problem-solving process. They were only partially successful in this, especially when the respective experts supported their solutions with AI-assisted analyses. AI support itself was considered a tool function in all teams. AI-supported results tended to be treated as particularly relevant or particularly reliable.

Figure 3: Comparison of working methods depending on the availability of AI support.
Figure 3: Comparison of working methods depending on the availability of AI support.

AI integration changes working methods and work results

The results of the four test runs in the human-AI role development laboratory show: Over the course of the simulation, the teams demonstrate progress in integrating AI into their work processes. They critically examine the AI outputs and achieve higher speeds with AI support.

However, they are unable to improve the quality of the key work results (health recommendations); in fact, there is a slight downward trend. This is related to the fact that working with AI support is accompanied by a change in the team’s problem-solving behavior. The teams are confronted with largely reliable findings but are unable to classify them in the same way as they did without AI support.

The results tie in with existing literature that reports negative effects of AI use on critical thinking [4, 11]. The present study shows that these effects can be observed within a few hours in a simulation setting. Via feedback on quality and the guided reflection round, however, this observation was accessible to team members in the simulation setting. It therefore stands to reason that an experience-based and workplace-oriented simulation approach can address both facets of AI competence: those that relate to AI operation and those that relate to critically reflective problem solving with and without AI.

The study does not allow for generalizable conclusions. This is due, on the one hand, to the very small number of test groups. On the other hand, numerous alternative explanations for the above results are conceivable. These include fatigue effects or reduced effort due to the absence of real consequences in terms of quality and learning effects in terms of speed. It also remains unclear to what extent the simulation has different effects on students and working professionals.

The diversity of the test groups (students, scientists, team from the healthcare industry) only made it possible to confirm the basic plausibility of the scenario across different professional backgrounds. Despite these limitations, the study suggests an approach to developing AI competence, the exact effect of which needs to be investigated in more comprehensive studies.

The present results make it clear that the development of AI competence for collective problem solving requires an approach that goes beyond the development of technical understanding. Responsible use of AI therefore includes, in particular, anticipating changes in thought processes with regard to problem solving when using AI and taking appropriate countermeasures. In addition to the approach presented in the Human-AI Role Development Lab, other experience-based approaches are conceivable that can be integrated into operational practice in the workplace.

This article was written as part of the HUMAINE research project, which was funded by the BMFTR funding program “Future of Value Creation – Research on Production, Services, and Work” (funding code: 02L19C200).


Bibliography

[1] Bankins, S.; Ocampo, A. C.; Marrone, M.; Restubog, S. L. D.; Woo, S. E.: A multilevel review of artificial intelligence in organizations: Implications for organizational behavior research and practice. In: Journal of organizational behavior 45 (2024) 2, pp. 159–182.
[2] Raisch, S.; Fomina, K.: Combining human and artificial intelligence: Hybrid problem-solving in organizations. In: Academy of Management Review 50 (2025) 2, pp. 441–464.
[3] Annapureddy, R.; Fornaroli, A.; Gatica-Perez, D.: Generative AI literacy: Twelve defining competencies. In: Digital Government: Research and Practice 6 (2025) 1, pp. 1–21.
[4] Lee, H.-P.; Sarkar, A.; Tankelevitch, L.; Drosos, I.; Rintel, S. et al.: The Impact of Generative AI on Critical Thinking: Self-Reported Reductions in Cognitive Effort and Confidence Effects From a Survey of Knowledge Workers. Proceedings of the 2025 CHI Conference on Human Factors in Computing Systems. Association for Computing Machinery 1121 (2025), pp. 1–22.
[5] Li, H.; Kim, S.: Developing AI literacy in HRD: Competencies, approaches, and implications. In: Human Resource Development International 27 (2024) 3, pp. 345–366.
[6] Bankins, S.; Formosa, P.: The ethical implications of artificial intelligence (AI) for meaningful work. In: Journal of Business Ethics 185 (2023) 4, pp. 725–740.
[7] Wilkens, U.; Lupp, D.; Langholf, V.: Configurations of human-centered AI at work: seven actor-structure engagements in organizations. In: Frontiers in Artificial Intelligence (2023) 6, 1272159. DOI: https://doi.org/10.3389/frai.2023.1272159.
[8] Bick, A.; Blandin, A.; Deming, D. J.: The rapid adoption of generative AI (No. w32966). National Bureau of Economic Research. 2024.
[9] Al Naqbi, H.; Bahroun, Z.; Ahmed, V.: Enhancing work productivity through generative artificial intelligence: A comprehensive literature review. In: Sustainability 16 (2024) 3, 1166.
[10] Rampelt, F.; Ruppert, R.; Schleiss, J.; Mah, D. K.; Bata, K.; Egloffstein, M.: How do AI educators use open educational resources?: A cross-sectoral case study on OER for AI education. In: Open Praxis 17 (2025) 1, pp. 46–63.
[11] Gerlich, M.: AI tools in society: Impacts on cognitive offloading and the future of critical thinking. In: Societies 15 (2025) 1, 6.
[12] Schuckart, A.: GenAI and Prompt Engineering: A Progressive Framework for Empowering the Workforce. In: Proceedings of the 29th European Conference on Pattern Languages of Programs, People, and Practices (2024), pp. 1–8.
[13] Johri, A.; Schleiss, J.; Ranade, N.: Lessons for GenAI Literacy from a Field Study of Human-GenAI Augmentation in the Workplace. In 2025 IEEE Global Engineering Education Conference (EDUCON) (2025), pp. 1–9.
[14] Kaspersen, M. H.; Musaeus, L. H.; Bilstrup, K. E. K.; Petersen, M. G.; Iversen, O. S. et al.: From primary education to premium workforce: Drawing on K-12 approaches for developing AI literacy. In: Proceedings of the 2024 CHI Conference on Human Factors in Computing Systems (2024), pp. 1–16.
[15] Creswell, J. W.; Clark, V. L. P.: Designing and conducting mixed methods research. New York 2017.

Your downloads


Potentials: Leadership Management Training

You might also be interested in

MAKI—A Digital Assistant for Practice-Based Learning

MAKI—A Digital Assistant for Practice-Based Learning

Why every factory is a learning factory
Olaf Resch ORCID Icon
With the help of digital assistants, academic teaching is possible in any factory. In order to achieve the best learning effects, however, the interests of all stakeholders must be taken into account. The factory wishes to deploy its employees quickly and productively, the learners desire a positive learning experience, and the educators want to illustrate abstract concepts in a meaningful and practical way. The only way to combine all of these perspectives is via a well-thought-out educational concept and highly functioning technology.
Industry 4.0 Science | Volume 42 | 2026 | Edition 2 | Pages 70-77
Serious Gaming and the Energy Transition

Serious Gaming and the Energy Transition

Collaborative knowledge generation and interactive understanding of complex interrelationships
Janine Gondolf ORCID Icon, Gert Mehlmann, Jörn Hartung, Bernd Schweinshaut, Anne Bauer
Conveying the complexity and multifaceted nature of the energy transition to a broad audience is a challenge. This article demonstrates how interactive serious games on a multitouch table can help make connections tangible and comprehensible. The games and the table were used in various conversational contexts. These are presented here in three case vignettes based on participant observation of the different applications, as well as situated and shared reflection. The vignettes demonstrate how interaction can trigger epistemic processes, enable shifts in perspective, and foster collective thinking, all of which are necessary for shaping the future of society as a whole.
Industry 4.0 Science | Volume 42 | 2026 | Edition 2 | Pages 62-69
Industrial Transformation via a Machining Learning Factory

Industrial Transformation via a Machining Learning Factory

A learning module to foster competencies for a sustainability-driven transformation
Oskay Ozen ORCID Icon, Victoria Breidling ORCID Icon, Stefan Seyfried ORCID Icon, Matthias Weigold
Sustainability-enhancing transformation processes are necessary in all sectors if we are to remain within planetary boundaries. This also applies to the industrial sector as a significant emitter of greenhouse gases. Employees need new competencies to master this complex task of industrial transformation. These range from CO2 equivalents accounting to the development and evaluation of transformation scenarios, including technical measures. The learning module developed here addresses these competency requirements and uses the example of the ETA factory to show how a competency-oriented learning module for industrial transformation can be structured. It essentially comprises four phases: data collection and CO2 equivalents accounting, cause analysis, development of measures and evaluation of measures.
Industry 4.0 Science | Volume 42 | Edition 2 | Pages 38-47 | DOI 10.30844/I4SE.26.2.38
Building the Future Workforce Today

Building the Future Workforce Today

Trendiation as a strategic framework for employee qualification and training
Jürgen Fritz, Sebastian Busse, Ingo Dieckmann, Torsten Laub
As Industry 4.0 and artificial intelligence reshape organizational capabilities, traditional training systems struggle to keep pace with evolving skill requirements. This paper introduces Trendiation—a structured methodology for translating emerging trends into actionable strategies—as a systematic approach to this challenge. Through a workshop-based application examining Edutainment, Human-Centered Design, and Workforce Transformation, we demonstrate how organizations can move from abstract trend identification to concrete qualification requirements and prioritized training initiatives. The method produces a traceable artifact chain spanning trend framing, capability-gap assessment, and implementation roadmaps. Participant evaluations indicate high perceived clarity and practical utility. By bridging foresight analysis with participatory design, Trendiation enables organizations to proactively cultivate adaptive capabilities and build learning cultures aligned with future work ...
Industry 4.0 Science | Volume 42 | 2026 | Edition 2 | Pages 22-29 | DOI 10.30844/I4SE.26.2.22
Data Quality and Domain Expertise for Resilient AI Deployment

Data Quality and Domain Expertise for Resilient AI Deployment

Integrating anomaly and label error detection in industry
Pavlos Rath-Manakidis, Henry Huick, Erdi Ünal, Björn Krämer ORCID Icon, Laurenz Wiskott ORCID Icon
AI implementation transforms work and worker-technology relationships in industrial quality control. This paper explores how approaches to data quality and model transparency support ethical AI deployment, fostering worker agency, trust, and sustainable work design in automatic surface inspection systems (ASIS). Recurring problems like data inefficiency, variable model confidence, and limited AI expertise point to key challenges of human-centered AI: user trust, agency and responsible data management. A solution co-developed with an ASIS supplier demonstrates that the challenges extend beyond the purely technical, underscoring the value of AI design that augments human capabilities. Technical solutions such as anomaly, label error, and domain drift detection are proposed to enhance data quality and model reliability. The insights emphasize the following generalizable strategies for resilient AI integration: understanding user-reported problems through a human-AI interaction lens, ...
Industry 4.0 Science | Volume 42 | Edition 1 | Pages 128-135 | DOI 10.30844/I4SE.26.1.120
Operationalizing Ethical AI with tachAId

Operationalizing Ethical AI with tachAId

Validating an interactive advisory tool in two manufacturing use cases
Pavlos Rath-Manakidis, Henry Huick, Björn Krämer ORCID Icon, Laurenz Wiskott ORCID Icon
Integrating artificial intelligence (AI) into workplace processes promises significant efficiency gains, yet organizations face numerous ethical challenges that stakeholders are often initially unaware of—from opacity in decision-making to algorithmic bias and premature automation risks. This paper presents the design and validation of tachAId, an interactive advisory tool aimed at embedding human-centered ethical considerations into the development of AI solutions. It reports on a validation study conducted across two distinct industrial AI applications with varying AI maturity. tachAId successfully directs attention to critical ethical considerations across the AI solution lifecycle that might be overlooked in technically-focused development. However, the findings also reveal a central tension: while effective in raising awareness, the tool’s non-linear design creates significant usability challenges, indicating a user preference for more structured, linear guidance, especially ...
Industry 4.0 Science | Volume 42 | 2026 | Edition 1 | Pages 50-59 | DOI 10.30844/I4SE.26.1.48