1 Introduction

Intelligence amplification (IA) aims at making people smarter with the use of artificial intelligence (AI) (Rouse & Spohrer, 2018; Schmidt, 2017a, 2017b; Xia & Maes, 2013). These AI technologies include image, speech and natural language processing, and problem solving and recommender systems that are more or less rule-based or machine learning systems (Rouse, 2020). Recent work in AI also has explored the idea of computational intelligence which is “… a set of nature-inspired computational methodologies and approaches to address complex problems associated with real world applications to which traditional methodologies and approaches are ineffective or infeasible. It primarily includes fuzzy logic systems, neural networks and evolutionary computation” (Sugumaran et al., 2017, p. 969). Many interesting technical developments in AI have been reported, some of the special issue of Information Systems Frontiers edited by Sugumaran et al. (2017) but the social perspective of organizational contexts and user-AI interactions that make these technological developments have value in practice have been as yet under-researched (Harper, 2019). Therefore, this article takes a social perspective on AI adoption, with a focus on developing knowledge to let human-AI interactions be effective, which we call an organizational learning perspective (Crossan et al., 1999).

In intelligence amplification, AI has an indirect effect on actual human decisions as a consequence of keeping the human in the loop of AI outcome quality checking, interpretation, and maintenance of these systems (Grønsund & Aanestad, 2020). Therefore, the actual AI effects on decisions are the outcomes of difficult to predict individual and organizational learning processes (Schmidt, 2017b) in interactions of human learning and machine learning, which Seidel et al. name triple loop learning (Seidel et al., 2018). A possible direct effect of artificial intelligence, where computers take over human decision making, is often infeasible in professional contexts because of the required decision accountability, the problem ambiguity, and the decisional uncertainty involved (Eberhardt et al., 2012; Etzioni & Etzioni, 2017). The simpler cases that are highly repetitive and involve not much diversity of sources and insights, could be run by rule-based or expert system, but many professional cases, like medical and legal use contexts, require an understanding of a personal situation and thus the application of contextually embedded knowledge “… grounded in the situated sociohistoric contexts of our lives and work” (Orlikowski, 2006, p. 460).

Artificial intelligence technologies may become supportive agents in an intelligence amplification socio-technical system involving one or more professional decision makers. Such a socio-technical system may support the process of understanding a case by the re-use and sharing of codified knowledge or the mining of new knowledge by data analytics (Markus, 2001). These socio-technical systems use and produce emerging knowledge because of ongoing learning processes (Nonaka, 1994; Orlikowski, 2006). This makes a static variance or causal analysis of the impact of intelligence amplification less relevant and requires the logic in such studies to be replaced by an emergent impact paradigm (Markus & Robey, 1988; Markus et al., 2002). This changes the assessment of intelligence amplification from researching the contribution of IA adoption (e.g., reduced mortality) as a dependent variable to the embedding of the technology with needs and the context of the stakeholders in a socio-technical system. Such embedding does not only require the understanding of its use for specific decision situations, i.e., a user-oriented learning process, but also the development of institutional environments as scaffolds of learning that both constrain and enable IA to be developed and implemented. These scaffolds “… include physical objects, linguistic systems, technological artifacts, spatial contexts, and institutional rules—all of which structure human activity by supporting and guiding it, while at the same time configuring and disciplining it” (Orlikowski, 2006, p. 462).

In more detail, we describe IA adoption challenges here for a clinical IA system for the purpose of learning about IA adoption as an organizational learning process. Such a clinical IA system contains multiple AI techniques, like visualisation (Turkay et al., 2014), zooming, sorting, and filtering of relevant patient data, images and video recordings (Korotkov & Garcia, 2012) and analysis of data from multiple sources like the patient’s medical record and genomic data for diagnosis and treatment selection (Raghupathi & Raghupathi, 2014). Natural language processing (NLP) can also be used as part of a clinical decision support system (CDSS) to extract the meaning from written notes in medical records (Auffray et al., 2016; Holzinger et al., 2008; Ivanović & Budimac, 2014). A CDSS thus is a collection of artificial intelligence systems for decision making in a clinical context. A CDSS like Watson for Oncology incorporates a form of prescriptive analytics by ranking treatment alternatives along predicted effectiveness for a given diagnosis with the support of mining knowledge from over 600 medical journals, hundreds of different medical data sources, and statistical evidence (Somashekhar et al., 2018). Such a CDSS may also be part of an intelligence amplification systems as a secondary knowledge mining system, i.e.,“…analysts attempt to extract knowledge from records that were collected by others, possibly unknown to the reuser, for very different purposes” (Markus, 2001, p. 71). Markus (2001) states that secondary knowledge mining requires extensive datamining skills, knowledge and experience, but many more capabilities must be developed to make this effective for intelligence amplification.

The literature on medical CDSS has reported on many challenges in realizing intelligence amplification. McNut et al. (2016) and Rumsfeld et al. (2016) mention that for the fields of oncology and cardiovascular diagnoses and treatments respectively, CDSS need sufficiently large and reliable datasets, which are not always available. For realizing such large data sets hospitals will have to share their data (Dinov, 2016; Peek et al., 2014; Salas-Vega et al., 2015). However, a lack of systems interoperability and different data taxonomies prevent inter-hospital data sharing (Asokan & Asokan, 2015; Raghupathi & Raghupathi, 2014). Medical practitioners are also concerned regarding the accuracy of classifications and predictions when the data sets are too small (Budhiraja et al., 2016; Geerts et al., 2016; Kruse et al., 2016; Szlezak et al., 2014) and the algorithms used are often non-transparent or incomprehensible (Eberhardt et al., 2012; Maia et al., 2017). Intransparency of algorithms may result in unacceptable feelings of loss of reasoning control for medical professionals (Eberhardt et al., 2012; Holden & Karsh, 2010; Maillet et al., 2015). Experiences with unreliable registrations in patient files also contribute to mistrust in CDSS (Salas-Vega et al., 2015). Many of the reasons for not trusting and resisting CDSS are thus not technical or psychological (i.e., the medical professionals risk perception) but are rooted in ethical, legal, and managerial intelligence amplification requirements being insufficiently met (Abouelmehdi et al., 2017; Andreu-Perez et al., 2015). Coping with these IA challenges thus requires specific capabilities, including personal skills and emergent organizational and inter-organizational capabilities and thus we explore IA adoption as an organizational learning process.

This article aims at theoretical generalization [i.e., theory formalization of grounded insights (Glaser & Strauss, 2009)] from a CDSS IA development case and by this create an organizational learning theoretical foundation for IA development in organizations. The kind of theory that we realize here is an analytic theory, which “… provides a description of the phenomena of interest, analysis of relationships among those constructs, the degree of generalizability in constructs and relationships and the boundaries within which relationships, and observations hold” (Gregor, 2006, p. 619). By this theory, researchers and practitioners will be able to disentangle the complexity of human-AI interactions in processes of IA implementation and will be able to design systems. Organizational learning has been described in its foundational literature as the creation of improvements (i.e., single loop learning) or innovations (i.e., double loop learning) and the creation of norms, rules and conditions by which these knowledge creation processes may be done best (so-called deutero learning or institutional learning) (Crossan, 2003; Visser, 2007; Wijnhoven, 2001). Such learning processes are emergent knowledge creation processes that transform intuitive or tacit insights into more explicit knowledge that can be combined to larger collections of explicit knowledge that next have to be integrated with everyday tacit understanding (Nonaka, 1994). Following the insight that we can know more than we can tell (Polanyi, 1966), the single loop, double loop and deutero learning processes start and end with tacit knowledge of individuals that may be shared in groups, networks or organizations. Clearly, only parts of tacit knowledge can be externalized, i.e., made person-independent and re-usable. In shared work practitioners contexts (Markus, 2001), much of the knowledge that cannot effectively be explicated, like specific person-dependent skills and experiences, can be registered and shared as part of the group’s transactive memory (Majchrzak et al., 2013). Such transactive memory thus does not codify and store the knowledge content itself but contains a shared memory for group members of who has what knowledge and information, enabling a shared division of cognitive labor (Brandon & Hollingshead, 2004). Alternatively, the explicit knowledge created by secondary knowledge miners may be materialized in databases, decisions rules and data mining pattern outcomes with different levels of technological readiness (Mankins, 2009; Rouse, 2020; Russell & Norvig, 2016).The explicit knowledge that is created in the process of realizing the organizational conditions for IA may be materialized in (privacy) laws, inter-organizational data sharing standards, and IT management policies. Externalization thus first places implicit knowledge on a human independent medium like an information system. Second, this explicit knowledge can be combined with other explicit knowledge, e.g., by the creation of data warehouses, analytics outcomes, and rule-based expert systems. Third, the more advanced combined knowledge can produce recommendations to decision makers that decision makers can internalize by integrating it with personal experiences, values, and skills, i.e., tacit knowledge. Finally, professionals may discuss problems and recommendations with colleagues for verification and to learn from each other or develop new standards and guidelines via a learning process called socialization. Socialization learning has group tacit knowledge and transactive memory as combined tacit knowledge outcomes. Similar knowledge emergent processes also exist for the deutero learning process that set the rules, plans and conditions for single or double loop learning in the IA context.

This brings us to the following research question:

What challenges exist in the socialization, externalization, combination, and internalization stages of intelligence amplification adoption?

The next section further describes the organizational learning approach to IA, after which we explain the case methodology used for detecting relevant adoption challenges. Section four presents our results of a medical IA adoption case and section five discusses the generalization of these results for other professional non-medical contexts. Lastly, the sixth section entails the conclusion and reflects about further research.

2 Intelligence Amplification Adoption as Organizational Learning

The information systems field has extensively discussed IT adoption. One approach here is the “theory of innovation adoption” which is grounded in Rogers (2010) work and describes innovation adoption as a multi-stage process of awareness development, persuasion, decision to adopt, implementation and continuation of use. These stages include learning processes where the person collects information first of where the innovation is about, next to decide about adopting, and finally collecting information for deciding on continuation of adoption and possible more extensive use. The innovation process affects five categories of adopters that represent different levels of technology adoption in the social system. These groups are named innovators, early adopters, early majority, late majority, and laggards, and they differ on their willingness and ability to take varying levels of risks with innovation adoption decisions. Innovators are willing to take risks, have the highest social status, have financial liquidity, and thus can absorb decision failures, and have closest contact to scientific sources and other innovators. Their risk tolerance allows them to adopt technologies that may ultimately fail. Laggards are the last to adopt an innovation. Individuals in this category show little to no opinion leadership. These individuals typically have an aversion to change-agents. Laggards typically tend to be focused on "traditions", lowest social status, lowest financial liquidity, oldest among adopters, and in contact with only family and close friends. In a social system, the distribution of members in each category may follow a bell shape with innovators and laggards being the smallest groups and the early majority and late majority being the largest groups. Clearly, innovation adoption is not a simplistic calculation of costs and benefits or the collection of information for making such a calculation. It also involves social pressure, norms and values, and personal risk and prestige. In AI adoption, some professionals may want to be innovators or early adopters for generating prestige where others will be laggards and risk averse. For being an innovator or early adopter, one needs to have the resources and reputation to be able to act as such (Burke et al., 2009).

Technology acceptance models have been developed to further identify personal or environmental factors that influence an individual’s adoption decisions. Venkatesh et al. (2003) have identified eight models of technology acceptance: the theory of reasoned action, technology acceptance model (TAM), motivation theory model, theory of planned behavior (TPB), a combined model of TAM and TPB, a model of PC utilization, innovation diffusion theory as grounded in Rogers’s work, and social cognitive theories of technology adoption. They combined these theories in a unified theory of acceptance and use of technology (UTAUT) (Venkatesh et al., 2003). The UTAUT holds that the key constructs that explain use intention and actual information technology use are (1) performance expectancy, (2) effort expectancy, (3) social influence, and (4) use facilitating conditions. Performance expectancy and effort expectancy are both used by people for making an innovation adoption cost benefit calculation. Innovators have higher performance expectations of innovations than laggards. Together with possible social influences like group pressures, norms and values, and active advertising, people will form a level of usage intention, which in its turn may influence actual usage behaviour. Facilitating conditions may be given to individuals for using a technology, like financial support, support expertise, a high-quality usage infrastructure, and thus also touches policies and infrastructures that are created and offered by the institutional contexts of potential adopters (Hull & Lio, 2006). This means that in contrast to expectations and social influence that influence an individual’s intention to use, poor facilitation may turn a person with a high intent to usage to an actual non-user. Gender, age, experience, and voluntariness of use are posited to moderate the impact of the four explanatory key constructs on use intention and behaviour.

Venkatesh et al. (2003) found in a longitudinal study that UTAUT accounts for 70% of the variance in behavioural intention to use IT and about 50% in actual use. More recent work of Parasuraman and Colby (2015) offers a technology readiness index which measures optimism and innovativeness as motivating predispositions for IT adoption, and discomfort and insecurity as inhibitors of IT adoption. Both these variance studies, however, do not explain how use intentions and actual use can be developed as part of an awareness creation, persuasion and motivation process for the creation of use intentions and actual use (Markus & Robey, 1988; Wheeler, 2002). AI adoption has many organizational consequences and therefore the development of knowledge for user adoption is a critical organizational learning process that requires socialization, externalization, combination, and integration of relevant knowledge for adoption decisions. For making the broader IA adoption decisions, the decision makers have to create knowledge of different kinds in multiple learning steps. First the decision maker has to become aware of possibilities and thus to externalize insights in explicit notions of performance and efforts for making a cost–benefit calculation (Rogers, 2010). Because decision making for IA adoption mostly requires the collaboration of work practitioners (Markus, 2001), the decision makers have to socialize and agree on values before such a costs benefit calculation can be made. The outcome of cost–benefit calculations has to be related with existing opportunities and constraints that are to some extent materialized in existing infrastructures, other collateral systems, IT policies and rules, resulting in a combination of explicit knowledge of several kinds. For the actual implementation of the decision, the system has to be internalized by changing existing ways of thinking and working of individual users. Especially for IA systems, the need for understanding AI outcomes and having AI well embedded in processes and people’s way of working is vital to let it work well as a decision support tool (Orlikowski, 2006). Finally, all people involved in use and support of the IA system will have to evaluate it effectiveness for realizing shared knowledge and beliefs to further use or extend its use by repeating the learning cycle starting at a more extended level of awareness. These learning cycles in the context of artificial intelligence use are not only human learning processes, but also involve machine learning, single loop and double loop, and the integration of machine learning outcomes with human learning outcomes, named triple loop learning (Seidel et al., 2018).

Because IA adoption is the adoption of AI technologies for professional decision making, IA adoption happens in and for an organization context. The learning processes therefore are organizational, and organizations may enable or constraint the IA adoption learning process in several ways. Organizations also may learn to influence the single loop, double loop and triple loop learning processes. These learning enabling and constraining processes are named deuteron learning (Visser, 2007; Wijnhoven, 2001). Deutero learning involves the development of dynamic capabilities, i.e., the continuous building up of resources for innovation (Kraaijenbrink et al., 2010). The essence of the dynamic capabilities (DCP) development is a continuous organizational effort of acquiring and reconfiguring capabilities for staying in tune with changing markets and technologies (Lavie, 2006). Similarly, intelligence amplification requires a capability of learning to find out how AI can be deployed effectively, i.e., the creation of norms, processes and scaffolds for individual organization members to learning to adopt AI and develop an IA socio-technical system. Such norms and processes are both enablers and constraints. Wheeler (2002) describes how IT related capabilities can be created via a Net-Enabled Business Innovation Cycle (NEBIC). This theory describes the cycle of value creation in net-enabled organizations through four dynamic capabilities: (1) choosing emerging/enabling technologies, (2) matching with economic opportunities, (3) executing business innovation for growth, and (4) assessing customer value. The relations between these capabilities are processes that describe learning from each of the four capabilities, communicating the results to the following capability, and feeding back market-based metrics. Translated to a more general theory of IA adoption, the choosing capability includes routines to create insights on emerging and enabling technologies that could support IA-enabled value creation. These insights could relate to, for example, mobile technologies, Internet technologies, and AI technologies and their technological readiness levels (TRL) (Mankins, 2009). The choosing capability has a broader focus than just technology. Choosing a technology goes together with selecting a technical platform and ecosystem and will be done at an IT policy level as well as the professionals view on its potential, like with organizational learning it involves the explication of feelings of usefulness for the organization. The resulting choosing insights are input to the matching capability, which includes routines to combine these insights with concrete business process activities and strategic goals of organizations. The matching capability, i.e., a combination learning activity, aims at revealing new IT-enabled economic opportunities. When these opportunities are detected, the next step is to internalize the detected opportunity insights by integrating them with existing processes and mindsets and finally to evaluate its success in terms of customer and organizational values, i.e., to create new group tacit knowledge. Each step requires a stage gate to be passed as an outcome of an organizational learning process performed by stakeholders with outputs consisting of individual’s intention to use and facilitating conditions that are created for actual usage.

The Nonaka model of organizational learning states that through socialization people share tacit knowledge and create a group level tacit knowledge. This group tacit knowledge contains facts and beliefs that people share but also norms, values, and standards. The externalization process transforms tacit knowledge into explicit and codified knowledge. This kind of knowledge is placed on a person independent “material” medium, like books, databases, manuals, and recipes. Via the combination process, content of these person-independent media can be combined to a larger set of combined knowledge. Such combined knowledge can consist of data warehouses and larger knowledge systems like digital encyclopaedias and recommender systems. The internalization process translates (group) explicit knowledge back into a personal understanding by integrating the explicit knowledge components, like recommendations, calculations, and reasoning, with what the individual person thinks is appropriate in the specific decision context. Thus, in internalization the explicit knowledge is evaluated on ethical, logical, and common-sense beliefs. Knowledge thus develops by a continuous emergent process of socialization, externalization, combination, and internalization. For IA adoption, this learning process thus is both user-oriented single, double, and triple loop learning and an institutional deutero learning process involving the development of enablers, like motivators, capabilities and resources, and constraints for these individual learning processes. Figure 1 gives a variant of Nonaka’s SECI model to summarize these learning processes as a system dynamic process, which means that knowledge stocks (the boxes in the model) receive growth from a learning process that delivers an inflow of new knowledge (wide arrows in Fig. 1). These inflows are more voluminous or speedily if they reuse insights from previous stocks represented by the thin broken link arrows that inserts the stock knowledge in the learning process and the deutero enabling or constraining variables (represented by the ovals in Fig. 1). (For more information about system dynamics modelling, the reader is referred to (Fortmann-Roe, 2014)). Note that we assume no unlearning as we have no outflows from the knowledge stocks. This means that we see the learning as continuously increasing the knowledge needed for adoption Also note that externalization involves the codification of existing knowledge, whereas combination may result in new insights, thus double loop learning. The triple loop learning, i.e., the learning from the AI systems and integrating these new insights in the stock of knowledge, only happens after the AI-based recommendations are delivered to the decision maker in the internalization process. In the socialization process, participants in this process may share existing (single loop learning) and new knowledge (double loop learning learning). The deutero learning only involves the influencing of these learning processes. Note that some authors have equated deutero learning with triple loop learning in the past, but we follow Seidel et al. (2018) who reserve the term triple loop learning for the integration of human and machine learning processes, and we reserve the term deutero learning for learning to manage the organizational learning processes (Visser, 2007; Wijnhoven, 2001).

Fig. 1
figure 1

The IA adoption learning process

3 Methodology

3.1 Case Study and Context

This research wants to describe the organizational learning challenges for a concrete implementation of an IA system and to generate a generalizable formal theory out of these grounded insights (Glaser & Strauss, 2009). To realize this, we use a rich case description in which key stakeholders were interviewed regarding their view on the adoption of an AI-based CDSS (Klein Koerkamp, 2019). This case is from a University Medical Centre, which initiated the “Analytics for Medicine” program in spring of 2017 to enable personalized healthcare with data analytics. This is a hospital-wide program with a special team of clinicians and data scientists. The case focuses on a CDSS pilot named Sepsis Identification Speed, which was run by the Neonatology department. Babies that are born too early are sensitive to infection. Treatments with invasive procedures, such as intravenous lines, blood samples and ventilation are all potential entry points for bacteria and add to the risk of illness. The Neonatology department wants to know as early as possible if a patient will become ill and what treatment is most appropriate. The current care process is as follows: (1) the physician suspects an infection, (2) the physician takes a blood culture, (3) this blood culture is examined for bacteria, and (4) when the culture is positive, the blood is examined by gram staining and bacteria are coloured to make them visible under the microscope to identify the species. This process can take up to 48 hours which can be crucial in the development of the infection and the administration of antibiotics.

The CDSS aims to support the physicians when they consider administering antibiotics. The CDSS focuses on predicting with a minimum of false negatives. False negatives are the most dangerous situations because a false-negative advises not to give antibiotics when it is needed. But to complicate the decisions, following false positives is harmful as well, because administering antibiotics has some negative side effects as well. The single loop learning process focusses on speeding up the predictions and decision process, where the double loop learning process aims at reducing prediction error. The CDSS uses different data sources from the database of neonatology which consists of 6000 children born between 24 and 32 weeks. This data originates from several systems whose data must be integrated and prepared within data management before analysis. For increasing its prediction reliability, the hospital develops cooperation with other hospitals to generate a larger dataset for analytics. The model development method applied for analysis is the ‘gradient boosting’ technique of predictive machine learning.

3.2 Data Collection

The units of analysis in this case study are the different CDSS project stakeholders’ opinions. Besides of the directly intended medical users, we identify political and financial stakeholders, IT support stakeholders, vendor developers, clinical ethicists, clinical methodologists, and regulator stakeholders. Based on this analysis, the respondents shown in Table 1 were selected as the units of observation in the case study. The respondents were directly approached, visited at an event, and selected and approached from the vendor’s network and advise from other interviewees. Some interviewees have overlapping stakeholder roles as can be seen in Table 1.

Table 1 Interviewee numbers, descriptions, and stakeholder roles

The case study used semi-structured interviews which provide openings for a narrative to unfold while also applying questions based on the literature review (Galetta, 2013). The unscripted narrative enabled the researcher to explore the respondents’ view. Interviewees were asked to describe features, components, processes, or activities of the CDSS in which barriers for implementation may occur. Moreover, after mentioning the barriers, interviewees were asked about possible ways to meet them in the shorter or longer run.

3.3 Data Analysis

The data from the semi-structured interviews have been recorded and transcribed. Subsequently, the transcribed statements have been coded with the ATLAS.ti qualitative data analysis software. To ensure the research reliability, the transcribed interview has been sent to the specific interviewee with the request to supply feedback and corrections where needed (Whittemore et al., 2001). The total meaning coverage of the used concepts [i.e., content validity (Babbie, 2015)] within the interviews is assessed by cross checking the concepts in the questionnaires with the healthcare & technology experts of the CDSS software vendor. The author of this article next classified relevant interview statements by organizational learning categories socialization, externalization, combination, and integration and in more depth selected indications of IA adoption challenges as deuteron learning challenges. The external validity includes the inductive generalization of the CDSS case by a comparison with different IA contexts to be analyzed in the section after the case results presentation. Generalization here is thus an inductive formalization of grounded insights (Glaser & Strauss, 2009).

4 Results: CDSS Adoption Challenges

This section presents data for the four organizational learning categories of the SECI sequence, which emphasizes the importance of tacit knowledge as the start and the end of any organizational learning. All externalizations and combinations that may follow from the learning process are reflections or extensions of tacit understandings but they are also inputs for new tacit understandings. For each SECI learning process, we describe the stakeholder view by presenting relevant statements and possible deuteron influencing variables.

4.1 Socialization Challenges

Socialization is the sharing of tacit knowledge among people. Tacit knowledge may be any insight or experience that is not formally defined and highly person-dependent. R7 (healthcare director vendor): “Hospitals have (…) a cultural problem, the physicians are too distant from the IT department.” Involving physicians within the development stage of the CDSS will be beneficial for trust and acceptancy of it. Therefore, R11 (ex-physician and data scientist) says: “It is important to bring along a group of physicians within this process or else you will get the 'not invented by me syndrome'.”

Including a regulatory expert in the R&D team might be beneficial to the CEFootnote 1 approval process preparation of the CDSS project team. R22 (Clinical CEO notified body) says: “A regulatory professional within a R&D team involved from step one can think of what Notified Body or food and drug administration (FDA) market approval conditions are”. According to R19 (inspector e-health), the EU is currently developing norms for artificial intelligence which are also applicable to CDSS. These norms could also give a greater legal framework. R22 (global clinical director notified body) says: "The inspection of healthcare and representatives of other EU member states validated the new (…) procedures during joined assessments. Very often during these joint assessments the EU member state representatives had different interpretations of the regulations. Hence, it took 7 years to write the new legislation, then you have a meeting with representatives of a notified body from several countries and the healthcare inspection and they are still discussing what the rules exactly say.”

Thus, the socialization process is not just of physicians and their teams, but also with technical, ethical, quality control (inspection) and legal stakeholders. All must understand each other. In this process mutual influencing is of course happening.

4.2 Externalization Challenges

Externalization is the transformation of tacit knowledge to explicit knowledge by codification of rules and knowledge and the registration of events and facts for managing them on person-independent media. The term “knowledge” is used here in a broad sense in the many ways that people say to know something, and thus includes facts and figures, theories and explanations, methods and skills, and experiences (Mingers, 2008).

Some interviewees say that the CDSS adoption is challenging because of several externalizations needed. For example, R5 (professor health informatics) says: “I believe that a lot of professionals are not satisfied with the registration load….”. R1 (ex-chairman board of the hospital board) says: “There is a tension between how we can standardize the systems input and how can we keep the input requirements user friendly”. Even when these data would be available and correct, a lot of data processing and analytics is needed which is not always compliant with CE quality norms. R11 (ex-physician and data scientist) says: “Some data producing machines within the hospital are validated and CE approved for research and not for healthcare, which is needed for CDSS, a new CE approval of the data warehouse is required to realize this transition.”

Physicians also hold professional expectations that make them hesitant to work with a CDSS. R20 (ex-physician and analytical entrepreneur) says: “CDSS technology enforces strict working according to guidelines and thus may deprive physicians from their sense of added value. This (…) makes physicians feel less valuated.” This is not necessarily a defensive attitude but could be based on sincere professional doubts. R17 (Business engineer and CDSS developer) says: “It is not clear if the improvement of care is worth the costs of data scientists, medical trial, infrastructure and maintenance of the CDSS project.”

A CDSS should use randomized control trial (RCT) outcomes, but R4 (clinical owner CDSS) says: “… randomized experiment where half will be exposed to the algorithm and the other half not (…) is seriously hard (…), we have to look at how we can show the clinical relevance without a randomized study.”

Thus, the externalizations needed for a realizing a working AI system are seen as risky in several ways and cannot easily convince medical professionals and hospital managers to choose in favour of them.

4.3 Combination Challenges

Combination is the process of bringing different externalized parts (like databases and applications) together in one larger system. R12 (clinical owner CDSS) says: “All the data the CDSS requires can be extracted from the hospital wide critical data layer for analytical processing by means of an API call. (…) This data layer is currently under construction, technologically feasible, however, realization depends on commitment and budget.”

Such a data layer would automatically collect, integrate and prepare the data from different data sources, but there is a lack of data sharing protocols. R8 (Senior technical consultant vendor) says: “Every system generates its own data in a way that is most easy for this system. Which is not wrong of course. Only when you want to join the data from these systems, you might realize that you should have done it in a different manner.”

….and R12 (clinical owner CDSS) says: “There is no automated process that collects data for analysis and makes it available to other solutions, all the data is still in its original source and has to be extracted and integrated manually by somebody to enable the following analysis.

Thus, the CDSS has to be viewed in the context of other data sources and systems and for this a well working policy, systems architecture, and standards have to be developed.

4.4 Internalization Challenge

Internalization is the process of humanly interpreting knowledge, insights and recommendations given by a system. For making recommendations acceptable for decision makers, they have to rely on transparent data analysis procedures. R10 (ex-physician and sales manager vendor) says: “The analytical process often requires transparency. (…) If a physician stands in front of a judge, he/she must be able to exactly explain how the process was executed.” ….. And the system must be compliant with the standards for CE approval. R12 (clinical owner CDSS) says: “However, the system still misclassifies some patients because they had other symptoms or something special, only 6 of the 500 children. This is not much, however, if all 6 babies die, this is hard to explain, of course.” And R13 (Program manager data analytics) says: “It does not matter if the algorithm says to intervene or not to intervene, the physician is always responsible and therefore is hesitant in using these systems.” Whereas the responsibility for medical decisions lies with the physician, there is a moral responsibility that lies with the developer. R16 (Ethicist and member of the medical ethical commission) says: “If an algorithm makes a mistake (…) the developer is partially responsible since s/he built this into the algorithm.”.

Thus, medical decisions in this case have high life impact and thus medical professionals only want to use the CDSS of its performance expectation is arguably very high.

5 Case Analysis

Figure 2 summarizes the CDSS development challenges mentioned in the case as deuteron learning constraints and enabler variables that influence the CDSS socialization, externalization, combination, and internalization processes. We further analyze the challenges for each of these processes while trying to generalize the insights of the case to general IA challenges per learning process. In a grounded theory development effort this is named the formalization of theory (Glaser & Strauss, 2009). We will do this by asking the question of what is specific for a health context case and thus also answer the question on what issues the CDSS case is not representative for many other IA adoption contexts. To develop a theory of IA adoption, we thus identify besides of the 4 explanatory variables of intention to use and actual use (Venkatesh et al., 2003) a sequence of organizational learning processes in terms of deutero learning variables, i.e., learning enabling and constraining conditions and contextual variables on which IA adoption projects may differ for their learning processes.

Fig. 2
figure 2

Intelligence amplification organizational learning challenges

The socialization process in which stakeholders of different disciplines develop a consensus on how to proceed in the IA project is critical for later IA adoption learning. In medical contexts, the knowledge input from lawyers regarding European and national laws on principles for privacy, data property, patient rights, safety, and quality of decisions explaining the need for their development, is a sufficient reason for medical professionals not (yet) to want to be engaged in the project, despite a vendor’s motivation to explain the opportunities. This also points at an important contextual factor for medical IA systems: law and regulation. Because medical professionals act on people’s life and health, a good legal context for this is probably more important than for other professions like education, software development, or architecture (Willmott et al., 2016).

In the process of further externalizing how the CDSS could look like, medical professionals may still have doubts regarding the quality and cost–benefit of such a system for medical professionals. This means that after the legal issues are settled, the system should be at the proper technological readiness level (Mankins, 2009). This also may require substantial testing before being considered for use and further development or implementation. Software testing is without doubt important for any information system, but for medical devices the technical readiness level must be very high because any risk taking in use may be a risk with again large human and legal consequences (Yusif et al., 2017).

If this externalization hurdle could be taken, the proposed system could be integrated with other systems and the broader organizational and inter-organizational landscape. For this, hospitals need sometimes extra budget and resources, which only can be acquired if the IA system meets a positive organizational cost–benefit calculation. Such cost–benefit calculation has to show items like time saved by professionals, increased hospitals incomes, or less costs of treatments. All these variables are often more difficult to assess for medical contexts (Henshall et al., 2012; Husereau et al., 2014). But a better treatment also has to cover its cost from an organizational perspective. So this makes IT investment decisions in the medical field sometimes very difficult (Castlen et al., 2019; Lakdawalla et al., 2017).

When the combination hurdle has been taken, the IA system can be implemented and given for use and the internalization hurdle comes next. For internalization, the medical professional must become convinced that recommendations from an AI system can be truly trusted and that the medical decision is thus better backed with state-of-the-art knowledge then without. Here, the medical field deviates again from other fields. Although any professional must be trustworthy, for a medical professional it is less easy to say that an erroneous decision was taken because a system recommended it (Eberhardt et al., 2012; Grote & Berens, 2020). This is especially the case when decisions are very complex, and an optimum choice option is hard to find. In these cases also the patient involvement in the decision making may be larger because patients have to bear the consequences of choices made (Charles et al., 1999). In such situations, an AI system may be less valuable, and in actual use the preferences, norms and values of stakeholders will determine much of the actual choice. Clearly, the medical field differs from most other professional disciplines by the large trust and the high personal responsibilities that need backing. Legal expert systems are especially well known for their rule-based character, as their knowledge is not probabilistic and cannot so easily be based on large numbers of near-equivalent cases (Leith, 2016). In the field of education, educational analytics are an upcoming field but still very new (Macfadyen, 2017). In engineering, rule-based systems may be more common as the natural scientific laws with which engineers work have a highly reliable and stable rule-based nature (Bullon et al., 2017; Savage et al., 2019) But also in these non-medical contexts, AI systems trust is an important value in IA for professionals.

If the internalization hurdle is taken, the learning process can continue to further steps, like the implementation of more features or the role out of the system to more professional users. At each moment that a hurdle cannot be taken, the process may go a step back, and the knowledge content and transactive memory that was build up can be usefully taken as a basis for further development.

6 Conclusions and Implications

This study realized an understanding of many challenges of artificial intelligence adoptions in the broader context of intelligence amplification development as an organizational learning process. More specifically, we described four organizational learning processes for IA adoption, starting with a tacit knowledge stock and finishing with an enhanced tacit knowledge stock by which this approach articulates IA as a socio-technical environment for making people become smarter. These learning outcomes can be incremental single loop improvements or more radical double loop learning innovations. The learning stocks do not only inform individual adoption decisions but also exist to realize the group learning and the organization-wide learning that are essential for letting AI systems deliver value. These learning processes are emergent and continuous, facilitated, speeded up or delayed by deutero learning variables. In our CDSS case, high requirements for legal backing, patient risk life avoidance, and technical-scientific maturity of the CDSS and its outcomes showed to be essential deuteron learning variables. Non-medical contexts may need less strict development on rules and laws, may be satisfied with systems of a lower technological readiness, even prototypes, may require less (inter-)organization wide arrangements because they can be used on an individual basis, and the professional decisions made with AI could be less affecting the rights and lives of individuals. These empirical propositions are in need of research in non-medical professional case contexts.

These insights have practical implications for the design of IA implementation projects. The project developer should assess the context, different learning stages and the thresholds that should be overcome per learning process. Given the complexities involved and the active involvement, contribution and commitments of multiple stakeholders, such projects should be management by more senior project managers who use the deutero learning variables for speeding up the process, moving the process to wanted directions, and allow profiting from new learning during the project’s execution (Pich et al., 2002). The need for learning during project executions has been articulated in modern agile project management methods before (Campanelli & Parreiras, 2015), but IA implementation projects demand also learning outside of the technical focus of a project. IA implementation project managers should be able to develop a development team as well as acquire financial and other resources from organization internal and external stakeholders. Project managers should be aware that the learning in IA implementations is not only human or technical single or double loop, but also triple loop, i.e., developing the interactions of AI learning and human learning with mutual influences (Seidel et al., 2018). This triple loop learning requires a careful gradual development of capabilities not only of computer scientists involved but of (professional) users as well (Grønsund & Aanestad, 2020). The end to such a learning process may be difficult to define as long as triple loop learning generates new insights. Besides of triple loop learning, the practice of IA implementations may also require the development of new organizational norms (i.e., deutero learning) that enable triple loop learning, e.g., the development of IT stakeholders’ responsibilities in decisions and the development of well working collaborative IT and user teams.

Theoretically and academically, this approach gives a revitalization of the importance of tacit knowledge stocks as the alpha and omega of organizational learning (Nonaka, 1994) and shows the relevance of this insight in the context of IA systems. This is in line with previous understanding of the difference between artificial or computational intelligence (Sugumaran et al., 2017) and intelligence amplification as a way of increasing human thinking power (single loop or double loop) with artificial intelligence as part of it (Berrar et al., 2013). However, more knowledge is needed of the diversity of organizational contexts and how these interact on the IA adoption process design and realization. For this, we suggest a number of research directions along different organizational learning processes.

Regarding deutero learning, one may for example ask the question if learning norms in other organizations than (academic) medical will be much different and what the influence of learning norms is on different challenges of IA implementation. Learning norms are part of the broader concept of organizational culture (Hofstede, 1980). Previous survey research on the impact of hospital organizational culture on IT adoption in general (do Carmo Caccia‐Bava et al., 2006) has identified that different corporate cultural elements like an organization’s internal versus external orientation and the level of flexibility or control influences and organization’s technological absorptive capacity and via that the success of IT adoptions in hospitals. We propose an update of this study by more depth towards the type of IT to be adopted, especially AI technologies, and the learning processes of users and other organizational stakeholders and with a comparison with other non-medical organizations, if survey data can be acquired. Also UTAUT survey studies (Venkatesh et al., 2003) can be extended and made more specific for IA implementation motivations and norms, including organizational learning as process and dependent variables. Such survey studies can be helpful in further understanding differences of IA implementations in contexts and give insights for project managers on how to identify, prioritize and handle the relevant deutero variables.

Regarding triple loop learning, the interactions of human users and AI systems can be studied as single loop or double loop learning processes. The difference between single loop and double learning is important and comparable to control and innovation (Argyris & Schon, 1978) or exploitation and exploration (March, 1991). Much previous research has been dedicated to the effectiveness of intelligent amplification for control (for example Althuizen et al., 2012; Bensoussan et al., 2009) but new research may be needed on using IA for innovation. Research may be useful to find out if IA can really make people smarter by generating new insights, new views, and creative solutions. Such research will have to go more in depth on creativity and how human creativity can be extended or improved with IA systems (Ardaiz-Villanueva et al., 2011; Rouse & Spohrer, 2018). Because creativity is often generated through interactions of multiple team members, one may also open research on triple loop learning as not only an individual human and system interaction, but also as interactions between multiple persons using multiple systems for innovation. In our theoretical framework this implies a focus on combination and socialization learning as part of IA implementation. Whereas combination has been studied before in the context of enterprise systems development (Strong & Volkoff, 2010), our case study demonstrates that combination is also a complex organizational process of developing standards and IT policies. Research on how socialization is both a start and end to IA implementation should answer research questions like how people develop motivations, norms and values for IA adoption and how new AI system usage influences people in their decisions and collaboration, and how people will maintain being the alpha and omega of knowledge in organizations.