Next Article in Journal
An Air Pollutant Forecast Correction Model Based on Ensemble Learning Algorithm
Next Article in Special Issue
Leading Role of E-Learning and Blockchain towards Privacy and Security Management: A Study of Electronics Manufacturing Firms
Previous Article in Journal
A Novel Denoising Algorithm Based on Wavelet and Non-Local Moment Mean Filtering
Previous Article in Special Issue
Private and Public Digitalization and Economic Resilience during COVID-19 Pandemic
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Ethical Impacts, Risks and Challenges of Artificial Intelligence Technologies in Business Consulting: A New Modelling Approach Based on Structural Equations

by
Andrea Gînguță
1,
Petru Ștefea
2,
Grațiela Georgiana Noja
3,* and
Valentin Partenie Munteanu
2
1
Doctoral School of Economics and Business Administration, West University of Timisoara, 300115 Timisoara, Romania
2
Department of Management, Faculty of Economics and Business Administration, West University of Timisoara, 300115 Timisoara, Romania
3
Department of Marketing and International Economic Relations, Faculty of Economics and Business Administration, West University of Timisoara, 300115 Timisoara, Romania
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(6), 1462; https://doi.org/10.3390/electronics12061462
Submission received: 2 March 2023 / Revised: 16 March 2023 / Accepted: 18 March 2023 / Published: 20 March 2023

Abstract

:
Artificial intelligence (AI) affects all aspects of a business, significantly contributing to problem-solving and introducing new operational processes within companies. Interest in AI is growing due to its capacities regarding the efficiency of operations, reduced working time, and quality improvements. However, to ensure proper development, businesses should also address the ethical effects generated by AI technologies. This research aims to identify AI’s ethical impact and associated challenges in the business consulting industry and the consultants’ perspective on AI’s future implementation in their specific sector. The methodology is based on the design of a structural equation modelling using data collected through a survey addressed to business consultants. The results highlight that ethical concerns are positively correlated with the identified harmful consequences of AI, such as high implementation costs, the possibility that this technology will lead to job losses, or a lack of human interaction and creativity. The consultants’ perspective on the future aspects of AI’s use in business consulting is negatively impacted by the following ethical outcomes: discrimination, invasions of privacy, denial of individual autonomy, unjustifiable results, and disintegration of social connection.

1. Introduction

Artificial intelligence is generally perceived as a research field of information technology and computer science, and it mainly focuses on designing and projecting intelligent providers [1], generating a remarkable impact on large domains of society and the economy [2]. The linguistic concept of “artificial intelligence (AI)” originated in 1956, and, as a field of study, it aims to conceptualize and represent intelligent behavior as a computing process. According to Boucher [3], AI illustrates a specific program that provides intelligent behavior by continuously examining its surroundings and making the necessary efforts to achieve its goal.
In terms of practical applications, AI has become one of the most attractive, but also emerging and disruptive technologies of the last years [4]. It is a widely debated research topic in various fields such as engineering, science, education, medicine, or economics. In business, AI-based systems with access to a vast database can be used in management, accounting, finance, human resources, marketing, and sales, potentially increasing revenue and decreasing costs. Artificial intelligence and the continuous learning capability of AI-equipped programs thus generate increased innovation, optimization within processes or resource management, and quality improvement [5].
In this complex framework, this research innovatively addresses the ethical impacts and associated challenges of AI in the business consulting industry and the consultants’ perspective on AI’s future implementation in their specific sector, through a new modelling approach based on structural/simultaneous equations. The main scientific goal of this study is to identify the ethical concerns and risks of AI related to business consulting and to determine the interlinkages between the consultants’ professional characteristics and field of activity, their perception of AI’s future application in business consulting, and their perceived detrimental consequences and ethical impact of these technologies. The research endeavour was planned extensively and fully to reach the main objective and then executed through phases, such as the outcome of one phase acting as the input for the next phase sequentially. More specifically, following the critical analysis of the specialized literature, we identified the advantages, disadvantages, and main ethical challenges associated with using AI in businesses. Next, based on these findings, we created a survey addressed to business consultants to determine their perspective on AI technology. Using these collected data from 98 respondents, we identify perceptions of AI adoption in the consulting profession and its associated risks and ethical challenges. Our methodology is based on a structural equation modelling (SEM) analysis of the interplay between the professional characteristics of business consultants and their perception of the AI risks and their willingness to accept this technology by deciding to invest in AI in order to implement it in their company or transfer a part of their tasks to an automatic program. SEM is an econometric analysis method that accurately determines the interdependencies between various constructs (i.e., independent, mediator, moderator, control, and dependent variables), which allows a richer understanding of their correlations [6]. The influences of some variables are thus identified through statistical and mathematical data leading to precise and highly accurate conclusions. Our results offer valuable insights for both practitioners and policy makers and bring to the fore the fact that ethical concerns are positively correlated with the identified harmful consequences of AI, such as the high implementation costs, the possibility that this technology will lead to job losses, or a lack of human interaction and creativity. Moreover, the consultants’ perspective on the future aspects of AI’s use in business consulting is negatively impacted by several ethical outcomes, such as discrimination, invasions of privacy, the denial of individual autonomy, unjustifiable results, and the disintegration of social connection.
The paper is structured as follows: The first part of this paper provides a brief presentation of AI and its professional and ethical impact on the business consulting industry. The second part is devoted to a detailed review of the scientific literature on the professional and ethical challenges of AI, while also grasping the working hypotheses. The third section embeds the research variables, the methodology applied, and the designed SEM model. The main results, discussion, and final remarks conclude the paper, complemented by Appendix A.

2. Professional and Ethical Challenges of Artificial Intelligence: Review of the Scientific Literature

2.1. The Evolution of Artificial Intelligence and Its Professional and Ethical Impact

An intelligent agent is defined as a knowledge-based system that analyzes the surroundings, reasons to interpret the perceptions, solves issues, and determines solutions to accomplish specific tasks for which it was designed [7]. It extracts its data and knowledge based on which it will act, and continuously adapts to the given assignment [7]. Artificial intelligence is intelligence associated with machines, in contrast to natural intelligence specific to humans and animals [8]. AI is mainly developed to have speech recognition, machine-learning, planning, and problem-solving attributes [9]. AI is applicable in computer science. Therefore, it implicates constructing devices that execute specific operations that would require intelligence if accomplished by humans [10].
The technical and managerial scientific literature offers multiple definitions of AI. Thus, AI can be seen as a new way computers are programmed to think in the same way that people do [11]. Artificial intelligence reflects the automation of human thinking, such as decision making, problem solving, or learning [12]. On the other hand, AI is characterized by a study of the computations that make perception, reason, and action possible [9]. Russell and Norvig [13] distinguish four approaches to AI that aim to simulate human thought, rational thinking, human actions, and rational actions by a programmable machine. Finally, it is reasonable to predict that AI will eventually impact all human activities, individual, professional, and social [14].
Large companies use this technology to implement marketing, human resources, or production strategies. However, increasingly frequent questions arise from the considerable development of AI applications and the determined implications, such as the replacement of the workforce and activities carried out by humans, or ethical issues: AI may cause significant job losses and could change the idea of employment [15], or it may exit from human control and even have the power of managing its own evolution [16].
Numerous researchers broadly study data privacy and security, as individuals should have complete control over their personal data, and their usage should not cause any harm or discrimination [17]. Privacy refers to controlling information about oneself and the right to keep it secret [18]. Artificial intelligence offers the ability to organize and store a large set of data, which entails the vulnerability and risk of personal information being accessed by other entities and used without the owners’ consent [19]. There may be situations where personal information is traded for a fee between entities and used in marketing and advertising processes, more quickly identifying the target market to promote their products or services [20].
Because controlling personal data is much more difficult online than in a physical format, most of the details of people’s lives are becoming increasingly accessible digitally, where data are collected and visible on high-capacity servers or in the cloud [21]. Many technologies that use AI amplify these problems. By using specific techniques, such as fingerprints or facial recognition, these technologies enable the identification of individuals and create a profile for each user [22]. Well-established legal protection of individual rights, including consumer rights or the responsibility for protecting intellectual property rights, is often lacking in digital products, or is challenging to implement [18]. Leslie [23] summarizes (Table 1) the potential damages a system based on artificial intelligence can produce.
The digitalization is continuously expanding, and technology is undergoing significant changes. Therefore, regulations should be adapted accordingly [24]. Artificial intelligence must comply with all applicable national and international legislation and regulations, and a set of requirements, such as safe, reliable, and robust algorithms to correct mistakes or inconsistencies throughout all phases of the AI systems lifecycle [25]. All AI systems should guarantee transparency, diversity, non-discrimination, and fairness while equally ensuring the accessibility of all users [26]. The European Commission’s High-Level Expert Group on AI [27] states that the well-being of society and the environment must be protected by intelligent systems and AI should be used to promote positive social change and improve environmental sustainability.

2.2. Business Consulting and the Impact of Artificial Intelligence on Business Professionals

Business consultants operate in multiple industries and with a variety of clients. Through their activities, they gather experience and valuable information that they can use and adapt according to the requirements. Nowadays, most consultants are asked to provide not only advice but also solutions, such as changing a company’s strategy [28]. Generally, there are no universal, standardized criteria for selecting a consultant because each client can define personal standards which reflect the company’s expectations and experience in consulting services. However, the price is often seen as an indicator of quality [29].
With digital access to data and equally available technologies, it takes more effort to differentiate between consulting firms. What can be considered a general characteristic is the consultant’s focus on the client and his needs, the goal of the consulting services being to solve the problems he faces as quickly and efficiently as possible [28]. Understanding and fully leveraging the data they operate with is one of the most important skills a consultant must possess [30]. Companies collect data continuously, being, at the same time, concerned with how these data are processed and exploited by consultants. They must maintain professional and ethical standards when working with their clients, having an obligation to keep the information obtained confidential [31].
The concept of digital business transformation is the use of technology to design unique business models, procedures, or techniques, providing greater efficiency, attracting additional revenue, and increasing the competitive advantage [32]. Digital transformation is also specific to companies in the field of business consulting, being able to have a positive and successful impact only with a solid strategy and management [33]. Machine-learning algorithms can build models and understand complex correlations through pattern detection, a challenging process for even the most promising and effective consulting teams [34]. If managed precisely, AI and automation can remarkably improve these firms’ functionality and customer services [35]. The benefits of using AI in business consulting are considered, according to Bayati [36], the following: AI has a fast and accurate ability to analyze large-scale data, better knowledge of the market and users, high efficiency in performing administrative tasks, and it can guide companies to allocate their financial resources properly.
All AI-based systems have a social and ethical impact on stakeholders and communities. The main goal regarding these systems is to achieve innovation to benefit society, but there are situations where they have the opposite impact. The new field of AI ethics has mainly appeared as a reaction to the individual and societal harms that the mishandling, poor configuration, or unintended damaging results of AI technologies can generate [37]. Leslie [23] suggests that, to develop and use a system based on artificial intelligence responsibly, it must be equitable and ethical, considering its impact on the well-being of individuals and the community. The use of an AI-based system must be fair and non-discriminatory, trustworthy, and transparent [23].

2.3. Research Hypotheses and Conceptual Model

Technology and specific computer programs have made employees’ work more efficient and, with AI’s development, these technologies’ limits continue to be exceeded. A computer could be programmed to analyze and enter data much faster than a person, but in order to follow these results, there remains the need for interpretation and creativity, which intelligent systems cannot yet replace [38]. To identify the perception of accounting professionals and educators on AI and its associated threats, Whitman [39] used survey data. In this study, younger and, therefore, less experienced participants believed that AI implementation could be helpful and improve their work by taking over their repetitive and administrative tasks [39].
The results of Abdullah’s [40] study indicated a need for training among healthcare employees regarding AI technologies and revealed that 78% of respondents worried that AI could completely replace the human workforce. Alexandre and Blanckaert [41] indicated that in the business consulting sector, small firms can hardly define and implement AI technologies, while bigger firms are able to internally develop and use it to make decisions. The implementation of AI is also strongly correlated with the company size and, at the same time, the opportunities for adopting AI programs are often found in more significant firms, while, for small companies, they are harder to implement [42]. Due to the financial capacity that allows large- and medium-scale firms to adopt the latest technology infrastructure, Gaafars’ [43] study revealed considerable distinctions between the responses concerning the application of AI tools according to the company’s characteristics.
Lestari and Djastuti [44] indicated different perspectives on AI in a specific sector (banking industry) according to the respondents’ attributes and company positions. Their research revealed that most employees working in frontline roles are concerned that AI technologies will be able to replace their jobs. On the contrary, respondents with back-office positions believed that human actions would still be required to conduct analysis procedures correctly and did not feel threatened by AI replacement [44]. These results show that professional characteristics (field of activity and position in the company) influence the perceived disadvantage of AI (possibility of replacing jobs). Taking into consideration the above literature, we assume that:
Hypothesis H1.
Professional characteristics and the field of activity determine the perceived disadvantages of using AI.
Many companies have used AI as an opportunity to develop, leading to increasing competition through consulting firms, including start-ups [41]. Assigning tasks to AI could determine that consultants would focus less on repetitive tasks and allocate more time to assessing problems and providing solutions. Workforce perspectives of the employees whose professional tasks are automated by a robotic process were examined by Zande [45] through eight interviews. The results showed that the interviewees positively perceived this technology because it lessened their workload.
Because many companies have already implemented AI systems, candidates capable of working with AI programs may be favored in hiring processes and a higher qualification would be required [41]. Understanding these innovations that fundamentally change a company’s internal processes can help leaders, managers, and the rest of the employees evolve. Choi [46] has shown that the clarity of the user and AI’s functions is positively associated with the user’s eagerness to accept AI implementation. According to Jaiswal [15], employees must possess five critical skills: data analysis, digital, complex cognitive, decision-making, and continuous learning skills.
Focusing on the internal implementation of AI, Alexandre and Blanckaert’s study [41] concluded that there are different results according to the firm’s size and its clients. Investments in AI technologies can have a notable effect on consulting firms. Besides the financial implications that require a concise strategy, the employees must also rapidly adapt to these changes [41]. All these arguments can lead to the following hypothesis:
Hypothesis H2.
Professional characteristics and the field of activity determine the future perspective of AI implementation.
Consultants spend considerable time analyzing data and information, an activity that a system based on AI could replace. According to Streib [47], the most common concern is that some professions and jobs will disappear due to the implementation of automatic analytics systems. Another issue pointed out refers to the explainability of AI processes [47]. For an AI management system to have interaction fairness while operating with employees and clients, the information it provides would require explainability of the procedures being carried out and its decisions or outcomes [48].
According to Ardon [49], AI mechanisms’ successful implementation must address workforce concerns. In his study, most employees are worried about job loss, unreliable algorithms, or security issues. Related to job losses is data privacy because AI systems will operate with more personal information. The capabilities of AI raised significant risks in business and the economy [47]. Streib [47] states that the insurance companies may use private data to evaluate the insurance premium or its coverage, and financial institutions could deliver credits and determine solutions based on such data.
Social interaction is a human need felt by every individual. Research has demonstrated that connections and interactions with colleagues in the workplace are essential to providing job satisfaction and reducing turnover intentions [48]. Therefore, the lack of communication could negatively affect a company’s professional activity and customers’ perceptions of that organization. The lack of social interaction could have an impact on the services provided by consultants, as clients could feel the absence of a real conversation. AI systems cannot offer the same advice or guidance as a person [38]. Consequently, we propose the following hypothesis:
Hypothesis H3.
The perceived disadvantages of using AI positively influences the perception of the ethical challenges of AI implementation.
AI and its capacity to develop patterns and generate insights might raise privacy threats, even without direct access to confidential data. Therefore, privacy and data security risks contest AI systems’ reliability. While there are doubts regarding the reliability of all technological devices, the machine-learning systems’ opacity and their unpredictability thus generate difficulties in testing their results [50]. The willingness to accept AI technology was negatively correlated to privacy concerns [46]. AI can affect ethical standards, safety, transparency, and public fear, and AI governance might be rigid and hard to control, discouraging new business ideas and containing their execution [47].
From a cognitive viewpoint, if employees have a better understanding of AI technologies, specifically their limitations and coverage, they can develop realistic expectations regarding these systems. Employees are less likely to pursue the suggestions of an AI management program if they are concerned that their actions will not have a beneficial effect or if they believe that it is not easy to use [48]. When encountering potential harmful outcomes emerging from AI, employees with higher levels of knowledge will be capable of determining solutions to overcome or diminish such threats [51]. Based on these findings, we assume that:
Hypothesis H4.
The perception of the ethical challenges of using AI negatively influences the future perspective of AI implementation.

3. Methodology

3.1. Research Variables

To conduct this study, we designed a survey containing questions and specific credentials from prior research validated in the scientific literature. The target group was Romanian business consultants, and the respondents were based in the cities of Timișoara and Arad. The questionnaire was distributed electronically over two months in 2022, and we received 98 complete responses.
Our main objectives were to determine the influences of AI on the business consulting industry regarding the risks and ethical challenges, as well as the business consultants’ perspective on AI implementation based on their professional characteristics and field of activity. We also studied whether the perception of AI’s ethical risks influences the business consultants’ future implementation perception of this technology. To measure their perspective on AI, we used the following variables (Table 2):

3.2. Methodology—Research Model

The research methodology is based on the application of structural equation modelling (SEM). SEM is a multivariate statistical method frequently used to conduct fundamental or applied research in behavioral, managerial, or quantitative social sciences [52]. It provides a complex underlying statistical theory and can address a variety of research hypotheses. As a modern econometric procedure, SEM demonstrates the capacity to estimate complex model correlations while measuring the error inherent in the indicators [53] and it has developed to support specific, accurate network models to match observations to theory [54]. Jeon [55] identified, among the advantages of SEM, the following: the possibility to capture latent variables and measurement equations; the capacity to determine the correlations between the dependent variables and pursue simultaneous estimation; the possibility to identify the direct effect, indirect effect, and total effect; to apply multiple statistical methods in one model (SEM contains measurement equations and structural equations); and to specify the reciprocal causal relationship between the latent variables. This method was also applied to previous studies regarding AI usage and employee perspective [56], and the impact of AI solutions and ethical concerns [57]. The popularity of this method can also be explained by the existence of various available and easily accessible software programs that allow researchers to apply SEM efficiently [55].
The SEM general diagram designed according to the above-formulated hypotheses is presented in Figure 1:
We used SEM to identify the relationships between our selected variables (Table 2) and our study focused on the influence professional characteristics and the field of activity have on the business consultants’ perception of the disadvantages of AI and their future perspective of AI implementation. We also analyzed the correlation between the perceived ethical challenges of AI and the identified drawbacks of this technology. Lastly, our study showed the impact of the ethical concerns evaluated by the business consultants on their perception of future AI implementation in consulting firms.

4. Results

The SEM model was processed through the maximum likelihood estimator (MLE) method in Stata (Figure 2). To test for the SEM models’ accuracy, we processed several robustness tests. In the goodness-of-fit test (Table A1), the comparative fit index (CFI) and Tucker–Lewis index (TLI) present a value close to 1, meaning that the model has a proper degree of fit. The coefficient of determination (CD) indicates that in over 54.9% of cases, the proposed observed variables influence the latent variables. We also processed the following tests: Cronbach’s alpha test (Table A2), where the total scale of the coefficient alpha is optimal (>0.7), and the Wald test (Table A3), that concluded with a p-value of 0.000 for one latent variable (CLG), 0.05 for the second (AIP), and 0.4 for the third latent variable (DSG).
In the present study, the perceived disadvantages of AI were considered on four dimensions: effects on communication (lack of human interaction caused by using AI systems, coefficient 1, the most decisive perceived disadvantage), social impact (job losses due to the automation of tasks, coefficient 0.84), economic–financial outcomes (high costs and significant investments for consulting firms in building and implementing AI, coefficient 0.35, the least relevant perceived disadvantage), and operational effect (the lack of creativity and innovation in AI programs, factor 0.92). The ethical challenges, in turn, were considered from five perspectives: the risk of discrimination (coefficient 1), the effect on personal responsibility and autonomy (coefficient 0.99), privacy invasion (coefficient 1.1, the most relevant perceived ethical risk), the potentially unreliable character (coefficient 0.9, the lowest perceived ethical risk), and the lack of social communication (coefficient 0.93).
The prospects for the future use of AI applications in consulting targeted four directions: the acceptability of AI implementation, measured by the consultants’ willingness to transfer a part of their tasks to automatic intelligent systems (coefficient 1), the awareness of the necessity for investments in AI programs (coefficient 1.1), the possibility of developing competitive advantages based on the use of AI technologies in the operational area (coefficient 1.1, the most relevant, along with the previous one), and the need of human personnel with previous experience in working with AI (coefficient 0.77, less relevant indicator).
Based on statistical analysis, we identified that hypothesis H1 and hypothesis H2 are only partially supported because of the reduced statistical significance of the estimated coefficients (Table 3). Our SEM results show a positive correlation between the professional characteristics and field of activity and the perceived disadvantages of AI (coef. = 0.23, Table A4) but present a reduced statistical significance (Table A4). Moreover, we identify that the professional characteristics and field of activity have a positive impact on business consultants’ future perceptions of AI implementation, but this hypothesis is only partially supported because of its statistical significance (Table A4).
Regarding hypothesis H3, the SEM estimation results present a positive correlation (statistically significant at the 0.1% threshold, Table A4) between the disadvantages of AI revealed by the business consultant’s evaluation and their perception of AI ethical risks in business consulting industry. The results show that the perceived disadvantages of AI are a significant positive predictor (coef. = 0.62, Table A4) of the respondents’ evaluation of the gravity and ethical impact this technology can have on their activity, such as discrimination, lack of responsibility, privacy invasions, and lack of human interaction and connection.
Regarding hypothesis H4, according to our study, the ethical challenges of AI are negatively correlated to the future perspective of AI implementation (statistically significant, p-value < 0.05, Table A4). The results show that the perceived ethical challenges of AI are a significant negative predictor (−0.28, Table A4) for the intention of business consultants to transfer a part of their job responsibilities to AI, to invest in AI professional application, to accept that AI could lead to a competitive advantage, or to prioritize candidates with AI knowledge in the recruitment process. The perceived ethical challenges strongly influence the perspective of materializing future implementation directions (coefficient 0.28) and the validation of this hypothesis means that the moral outcomes of using AI in business consulting also influence technological, economic, and strategic views. AI can be seen as a technology that can raise several ethical issues in consulting, such as algorithmic bias and the security of client data and information: the stronger these risks are perceived, the greater the potential future benefits are diminished.

5. Discussion

The previously described results are consistent with prior studies, but we could also identify some results contrary to our findings. Araujo [58] also underlined that privacy risks are negatively correlated with perceptions of fairness and the usefulness of AI capabilities, and positively linked with perceived threats. Christensen et al. [29] state that the focus placed on internal adaptation to the radical changes generated by the digital growth within consulting services is decisive, with consulting firms having to advise client companies pleasingly, while also considering cyber security threats. Reddy and Reinartz [59] also evidenced, among the AI challenges, the possibility of automation replacing the human workforce, the risk of customers losing their data confidentiality, and the rapid appearance of new competitors on the market.
Employee perspectives may intensely impact technology endorsement decisions, affecting a company’s innovation outcomes and performance [60]. Therefore, business consultants are less likely to transfer their administrative tasks to an automatic machine or to consider an investment in such technologies if they perceive these AI risks to have a critical impact on their activity. Choi [46] presented similar results and indicated that trust positively impacts the relationship between ability and the employees’ acceptance of AI technology. Therefore, a lack of confidence will cause employees to reject AI programs.
Firms are using AI tools to recruit, select, and manage their employees, and even if AI may seem neutral and objective, these systems function according to how they were programmed [61]. Indicating the risk that AI might replicate human biases or even create other types of discrimination, Kim and Bodie [61] refer to the case of Amazon. The company designed an algorithm to detect viable candidates for specific jobs but rejected the method because it devalued qualified female applicants. Prior studies also reveal that hiring processes and decisions are seen as less fair when accomplished by AI than by human personnel [62].
Previous research has also highlighted that employees are worried about the potential job losses AI adoption can cause [49,63]. On the other hand, Ourdedine [64] states that even if this technology will eventually replace low-skilled jobs, it can also be capable of creating new jobs as it develops and becomes more implemented through organizations. The lack of human interaction can also be distinguished in other studies [62], where AI interviews were generally perceived as less procedural and interactional just in comparison with the human conducted interviews.
Our result is also consistent with other previous similar studies [65,66], which indicated that the positive relationship between firm size and AI implementation is not supported at a statistically significant level. On the contrary, Pan et al. [67] indicated that the industry and the company size have no significant effect on AI usage. These results can be explained by the rapid appearance of new start-ups on the market that use AI or innovative technologies, regardless of their company size and professional characteristics.
However, referring to professional experience, our findings are inconsistent with Kambur et al. [68]’s study, where employees revealed different perceptions of AI based on their working experience. The previously mentioned research was conducted on managers and employees of the biggest firms in Turkey based on their capital. These results may differ in our study because of the domain and location of our target group.
The validation of the hypothesis that there is a direct relationship (coefficient 0.62) between the perceived disadvantages of using AI and the ethical challenges generated by them has several explanations in the consulting field. The most critical link is revealed at the level of affecting direct interpersonal interactions, which are still considered necessary in business consulting because dialogue and informal meetings can strengthen professional client–consultant relationships. Moreover, the use of AI in the consulting process may involve the collection and storage of clients’ personal data, equivalent to the violation of confidentiality and the risk of exposing some of their private information. At the same time, in many cases, AI algorithms make decisions without the consultants being able to understand the decision-making process, verify the findings, and are able to explain them later to their clients. Following the lack of creativity and innovation, AI can make wrong decisions, leading to adverse outcomes on the quality of the consulting services and amplifying the costs of development and implementation (this, however, being the least perceived disadvantage). In addition, some jobs (routine ones, especially, e.g., accounting and financial analysis) can be automated through AI, leading to job losses for professionals in these categories.
This research contributes to a better understanding of the effects of AI technologies in the business consulting sector, one of the business sectors in which AI will likely generate a significant impact in the future [69]. The novelty of the study compared to the existing literature is represented by the innovative configuration of a structural equation model based on four research hypotheses formulated through the analysis of the scientific literature and tested with an original questionnaire-type research instrument. Furthermore, the originality of this research consists in designing a new advanced modelling approach based on the authors’ visualization and construction of 17 observed variables divided into four latent variables, which aim to emphasize the business consultants’ perception of AI and its related drawbacks and ethical concerns. The results determine the general and specific implications for the consultancy companies’ management as well as for independent consultants.

6. Conclusions

The current research provides valuable insights into the ethical impacts, risks, and challenges of artificial intelligence technologies in business consulting. Our results indicate a partially supported connection between the field of consulting activity, defined by previous professional experiences, and AI disadvantages or future perspectives of implementing AI technologies. These findings reveal that the consultants’ professional characteristics and the companies’ dimensional aspects have a modest influence on how AI technologies can be implemented, generating opportunities or, on the contrary, disadvantages and risks. On the other hand, the results highlight statistically significant links between the perceived disadvantages, ethical challenges, and future implementation perspectives, with the ethical concerns acting as a mediating factor. If the perceived liabilities of using AI are more significant, ethical challenges are more relevant, and future implementation perspectives are negatively influenced. The discussions below focus on how the perceived disadvantages and underlying ethical challenges impact the prospects for the future implementation of these technologies.
Based on our findings, the perception of AI disadvantages strongly influences the ethical challenges of using AI. According to the business consultants’ perspective, our results show that the lack of human interaction, the possibility of replacing jobs, high implementation costs, and the lack of creativity of AI systems can lead to discrimination, denial of individual autonomy, privacy threats, unexplainable results, or disintegration of social connection in the business consulting industry.
These ethical concerns also determine the inhibition of the future implementation trend of AI. Business consultants are more willing to transfer their administrative tasks to AI and invest in new technologies if they perceive it to have lower risks with regard to their professional activity. The outcomes of these technologies also affect their perception of the competitive advantage gained by the consulting firms that use AI and the recruitment advantage some candidates possess because of their AI general knowledge.

6.1. Empirical and Managerial Implications

Overall, the implications of our study for researchers, practitioners, and consulting firms are associated with the managers’ and consultants’ understanding of the outcomes generated by introducing AI technologies in business consulting. Based on the results of the present study, it is recommended that organizations active in the consulting sector, as well as independent professionals, consider all relevant aspects determined by AI implementation in professional activities, including the ethical risks. These are determined by the perceived disadvantages of using AI, and they determine, in turn, a tendency for a more difficult acceptance of these technologies in professional processes and procedures. It can even represent resistance to change. To modify this tendency, it is necessary to adjust the perceptions regarding the reduction of human interaction, the risks of losing human relevance in consulting jobs, and the reduced creativity of AI. Adjusting the perception of the disadvantages of AI can also determine a reduction of perceived ethical risks related to possible discrimination, violations of private space and confidentiality, or professional autonomy. This process is also conditioned by exogenous causes, primarily determined by strict legal regulations regarding data protection or technological developments that avoid using discriminatory criteria in various professional practices.

6.2. Theoretical Implications

Digital transformation and emerging digital technology significantly impact numerous aspects of social and professional life in both technical and socio-economic fields (psychology, management, and accounting) and medical and educational areas. The literature analysis on the impact of AI usage reveals relatively divergent results regarding its advantages and disadvantages in various fields of activity. The theoretical implications of this research are represented by the continuity of study regarding the impact of AI in business and the effects of the employees’ perspective regarding ethical concerns on the future acceptance and usage of AI in work-related assignments/professional activities. The paper’s originality consists in developing and testing using SEM, a conceptual model that reveals a cause-and-effect relationship between the perceived weaknesses and ethical risks and limits of the perspectives of using AI in professional business consulting.

6.3. Limitations and Further Research

Subsequent studies and future research should concentrate on determining the factors that can reduce the perception of the disadvantages and ethical risks of using AI in this field or related areas (i.e., human resource management or project management) so that emerging technologies represent sources of competitive advantage rather than professional anxiety. The current research has some limitations induced mainly by a relatively reduced sample of respondents that are less dispersed geographically. Therefore, future research aims to capture the perceptions of business consultants from various European countries, in a comparative approach.

Author Contributions

Conceptualization, A.G., V.P.M. and P.Ș.; methodology, A.G., V.P.M. and G.G.N.; software, A.G. and G.G.N.; validation, A.G., V.P.M. and G.G.N.; formal analysis, A.G. and V.P.M.; investigation, A.G., V.P.M., G.G.N. and P.Ș.; resources, G.G.N. and P.Ș.; data curation, A.G. and V.P.M. writing—original draft preparation, A.G. and V.P.M.; writing—review and editing, A.G., V.P.M., G.G.N. and P.Ș.; visualization, A.G.; supervision, V.P.M. and P.Ș.; project administration, A.G. and P.Ș. All authors have read and agreed to the published version of the manuscript.

Funding

This research received funding from the West University of Timisoara.

Data Availability Statement

The dataset supporting the results reported in this article can be made available by the authors upon request.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. Goodness-of-fit tests.
Table A1. Goodness-of-fit tests.
SpecificationsSEM Model
Likelihood ratio
  Chi2_ms (115)168.891
  p > chi20.001
  chi2_bs (136)620.415
  p > chi20.000
Information criteria
  AIC4881.493
  BIC5024.225
Baseline comparison
  CFI0.889
  TLI0.868
Size of the residuals
  SRMR0.075
  CD0.549
Source: Own contribution in Stata 13.
Table A2. Cronbach’s alpha test.
Table A2. Cronbach’s alpha test.
ItemItem–Test
Correlation
Item–Rest CorrelationAverage Inter-Item CorrelationAlpha
Domain0.13100.00930.20950.8092
Experience0.33290.21840.19570.7956
Clients0.26970.15180.20000.8000
Interaction0.66020.58260.17330.7703
Jobs0.62600.54290.17560.7732
Costs0.36340.25100.19360.7934
Creativity0.65330.57460.17370.7709
Discrimination0.68140.60740.17180.7685
Responsibility0.62860.54590.17540.7729
Privacy0.69300.62100.17100.7675
Results0.67330.59790.17240.7692
Connection0.72590.65990.16880.7646
Tasks0.23900.11980.20210.8021
Transfer0.38130.27030.19240.7921
Investment0.40000.29040.19110.7908
Competitive0.34350.22980.19490.7949
Recruitment0.40920.30040.19050.7901
Test scale 0.18540.7946
Source: Own contribution in Stata 13.
Table A3. Wald test.
Table A3. Wald test.
VariablesChi2dfp-Value
AIP5.9820.0503
DSG0.6710.4143
ECLG31.8710.0000
Source: Own contribution in Stata 13.
Table A4. SEM detailed results.
Table A4. SEM detailed results.
Estimates, p-Value, and Standard Errors
AIP
CLG−0.279 * (0.115)
PCFA0.270 (0.177)
DSG
PCFA0.233 (0.285)
CLG
DSG0.622 *** (0.110)
Domain
PCFA1 (.)
_cons5.040 *** (0.165)
Experience
PCFA−0.414 (0.264)
_cons2.364 *** (0.134)
Clients
PCFA0.179 (0.190)
_cons2.970 *** (0.0942)
Tasks
PCFA0.201 (0.176)
_cons3.374 *** (0.0840)
Transfer
AIP1 (.)
_cons3.010 *** (0.0976)
Investment
AIP1.065 *** (0.206)
_cons2.939 *** (0.0979)
Competitive
AIP1.051 *** (0.246)
_cons2.929 *** (0.0963)
Recruitment
AIP0.773 *** (0.211)
_cons2.909 *** (0.0960)
Interaction
DSG1 (.)
_cons3.424 *** (0.133)
Jobs
DSG0.843 *** (0.0985)
_cons3.535 *** (0.114)
Costs
DSG0.354 *** (0.0959)
_cons2.707 *** (0.100)
Creativity
DSG0.924 *** (0.127)
_cons3.384 *** (0.137)
Discrimination
CLG1 (.)
_cons3.071 *** (0.124)
Responsability
CLG0.985 *** (0.139)
_cons3.071 *** (0.123)
Privacy
CLG1.091 *** (0.156)
_cons3.141 *** (0.134)
Results
CLG0.899 *** (0.147)
_cons3.030 *** (0.122)
Connection
CLG0.934 *** (0.161)
_cons3.263 *** (0.132)
var(e.Domain)
_cons1.629 * (0.806)
var(e.Experience)
_cons1.582 *** (0.257)
var(e.Clients)
_cons0.843 *** (0.128)
var(e.Tasks)
_cons0.655 *** (0.103)
var(e.Transfer)
_cons0.542 *** (0.109)
var(e.Investment)
_cons0.495 *** (0.107)
var(e.Competitive)
_cons0.475 *** (0.107)
var(e.Recruitment)
_cons0.674 *** (0.112)
var(e.Interaction)
_cons0.558 *** (0.125)
var(e.Jobs)
_cons0.440 *** (0.0923)
var(e.Costs)
_cons0.847 *** (0.124)
var(e.Creativity)
_cons0.845 *** (0.154)
var(e.Discrimination)
_cons0.681 *** (0.124)
var(e.Responsability)
_cons0.685 *** (0.123)
var(e.Privacy)
_cons0.779 *** (0.140)
var(e.Results)
_cons0.786 *** (0.132)
var(e.Connection)
_cons0.996 *** (0.165)
var(e.AIP)
_cons0.279 * (0.110)
var(e.DSG)
_cons1.123 *** (0.257)
var(e.CLG)
_cons0.383 *** (0.115)
var(PCFA)
_cons1.076 (0.831)
N98
Note: “Standard errors in parentheses, * p < 0.05, ** p < 0.01, and *** p < 0.001”. Source: Own contribution in Stata 13.

References

  1. Ndrejaj, A.; Ali, M. Artificial Intelligence Governance: A Study on the Ethical and Security Issues that Arise. In Proceedings of the 2022 International Conference on Computing, Electronics & Communications Engineering (iCCECE), Southend, UK, 17–18 August 2022; pp. 104–111. [Google Scholar]
  2. Saghiri, A.M.; Vahidipour, S.M.; Jabbarpour, M.R.; Sookhak, M.; Forestiero, A. A survey of artificial intelligence challenges: Analyzing the definitions, relationships, and evolutions. Appl. Sci. 2022, 12, 4054. [Google Scholar] [CrossRef]
  3. Boucher, P. Artificial Intelligence: How Does It Work, Why Does It Matter, and What Can We Do about It? Panel for the Future of Science and Technology, European Parliamentary Research Service. 2020. Available online: https://www.europarl.europa.eu/stoa/en/document/EPRS_STU(2020)641547 (accessed on 17 January 2023).
  4. Yigitcanlar, T.; Desouza, K.C.; Butler, L.; Roozkhosh, F. Contributions and risks of artificial intelligence (AI) in building smarter cities: Insights from a systematic review of the literature. Energies 2020, 13, 1473. [Google Scholar] [CrossRef] [Green Version]
  5. Cioffi, R.; Travaglioni, M.; Piscitelli, G.; Petrillo, A.; De Felice, F. Artificial Intelligence and Machine Learning Applications in Smart Production: Progress, Trends, and Directions. Sustainability 2020, 12, 492. [Google Scholar] [CrossRef] [Green Version]
  6. Albahri, A.S.; Alnoor, A.; Zaidan, A.A.; Albahri, O.S.; Hameed, H.; Zaidan, B.B.; Peh, S.S.; Zain, A.B.; Siraj, S.B.; Masnan, A.H.B.; et al. Hybrid artificial neural network and structural equation modelling techniques: A survey. Complex Intell. Syst. 2022, 8, 1781–1801. [Google Scholar] [CrossRef] [PubMed]
  7. Dzițac, I. Inteligența Artificială; Publisher UAV: Arad, Romania, 2008; pp. 28–31. [Google Scholar]
  8. Pan, Y. Heading toward Artificial Intelligence 2.0. Engineering 2016, 2, 409–413. [Google Scholar] [CrossRef]
  9. Joiner, I.A. Artificial Intelligence: AI is Nearby. In Emerging Library Technologies; Chandos Publishing: Amsterdam, The Netherlands, 2018; pp. 1–22. [Google Scholar]
  10. Fjelland, R. Why general artificial intelligence will not be realized. Humanit. Soc. Sci. Commun. 2020, 7, 10. [Google Scholar] [CrossRef]
  11. Sarker, I.H. AI-Based Modeling: Techniques, Applications and Research Issues Towards Automation, Intelligent and Smart Systems. SN Comput. Sci. 2022, 3, 158. [Google Scholar] [CrossRef] [PubMed]
  12. Fenwick, A.; Molnar, G. The importance of humanizing AI: Using a behavioral lens to bridge the gaps between humans and machines. Discov. Artif. Int. 2022, 2, 14. [Google Scholar] [CrossRef]
  13. Russel, S.; Norvig, P. Artificial Intelligence A Modern Approach, 3rd ed.; Pearson-Prentice Hall: Hoboken, NJ, USA, 2010; pp. 1–5. [Google Scholar]
  14. Makridakis, S. The forthcoming Artificial Intelligence (AI) revolution: Its impact on society and firms. Futures 2017, 90, 46–60. [Google Scholar] [CrossRef]
  15. Jaiswal, A.; Arun, C.J.; Varma, A. Rebooting employees: Upskilling for artificial intelligence in multinational corporations. Int. J. Hum. Res. 2022, 33, 1179–1208. [Google Scholar] [CrossRef]
  16. Bostrom, N. Ethical issues in advanced artificial intelligence. In Science Fiction and Philosophy: From Time Travel to Superintelligence; Schneider, S., Ed.; John Wiley & Sons: Hoboken, NJ, USA, 2009; pp. 277–284. Available online: https://nickbostrom.com/ethics/ai.pdf (accessed on 20 January 2023).
  17. Tucker, C. Privacy, Algorithms, and Artificial Intelligence. In The Economics of Artificial Intelligence: An Agenda; Agrawal, A., Gans, J., Goldfarb, A., Eds.; University of Chicago Press: Chicago, IL, USA, 2019; Volume 17, pp. 423–437. Available online: http://www.nber.org/chapters/c14011 (accessed on 10 January 2023).
  18. Müller, V.C. Ethics of artificial intelligence and robotics. In The Stanford Encyclopedia of Philosophy; Zalta, E.N., Ed.; Metaphysics Research Lab Stanford University: Stanford, CA, USA, 2020; Available online: https://plato.stanford.edu/entries/ethics-ai/ (accessed on 18 January 2023).
  19. Murdoch, B. Privacy and artificial intelligence: Challenges for protecting health information in a new era. BMC Med. Ethics 2021, 22, 122. [Google Scholar] [CrossRef]
  20. Costa, E.; Halpern, D. The Behavioural Science of Online Harm and Manipulation, and What to Do About It; The Behavioural Insights Ltd.: London, UK, 2019; pp. 13–30. [Google Scholar]
  21. Abdulsalam, Y.S.; Hedabou, M. Security and Privacy in Cloud Computing: Technical Review. Future Internet 2022, 14, 11. [Google Scholar] [CrossRef]
  22. Mann, M.; Smith, M. Automated facial recognition technology: Recent developments and approaches to oversight. UNSW Law J. 2017, 40, 121–145. [Google Scholar] [CrossRef]
  23. Leslie, D. Understanding Artificial Intelligence Ethics and Safety—A Guide for the Responsible Design and Implementation of AI Systems in the Public Sector; The Alan Turing Institute: London, UK, 2019; pp. 3–23. [Google Scholar]
  24. Forradellas, R.F.R.; Gallastegui, L.M.G. Digital Transformation and Artificial Intelligence Applied to Business: Legal Regulations, Economic Impact and Perspective. Laws 2021, 10, 70. [Google Scholar] [CrossRef]
  25. Freiman, O. Making sense of the conceptual nonsense ‘trustworthy AI’. AI Ethics 2022. [Google Scholar] [CrossRef]
  26. Felzmann, H.; Villaronga, E.F.; Lutz, C.; Tamo-Larrieux, A. Transparency you can trust: Transparency requirements for artificial intelligence between legal norms and contextual concerns. Big Data Soc. 2019, 6, 1–14. [Google Scholar] [CrossRef]
  27. European Commission’s High-Level Expert Group on AI. Ethics Guidelines for Trustworthy AI; European Commission: Brussels, Belgium, 2019; pp. 9–24. Available online: https://digital-strategy.ec.europa.eu/en/policies/expert-group-ai (accessed on 13 January 2023).
  28. Toppin, G.; Czerniawska, F. Business Consulting—A Guide to How It Works and How to Make It Work; The Economist: London, UK, 2005; pp. 3–13. [Google Scholar]
  29. Christensen, C.M.; Wang, D.; Bever, D. Consulting on the Cusp of Disruption. Harv. Bus. Rev. 2013, 10, 1–10. [Google Scholar]
  30. Banai, M.; Tulimieri, P. Knowledge, skills and personality of the effective business consultant. J. Manag. Develop. 2013, 32, 886–900. [Google Scholar] [CrossRef]
  31. Mingaleva, Z. Ethical Principles in Consulting. Procedia Soc. Behav. Sci. 2013, 84, 1740–1744. [Google Scholar] [CrossRef] [Green Version]
  32. Schwertner, K. Digital transformation of business. Trakia J. Sci. 2017, 15, 388–393. [Google Scholar] [CrossRef]
  33. Crișan, E.L.; Stanca, L. The digital transformation of management consulting companies: A qualitative comparative analysis of Romanian industry. Inf. Syst. E-Busin. Manag. 2021, 19, 1143–1173. [Google Scholar] [CrossRef]
  34. Libert, B.; Beck, M. AI may soon replace even the most elite consultants. Harv. Bus. Rev. 2017, 7. Available online: https://hbr.org/2017/07/ai-may-soon-replace-even-the-most-elite-consultants (accessed on 19 January 2023).
  35. Carmona, D. The AI Organization; O’Reilly: Sebastopol, CA, USA, 2020; pp. 80–82. [Google Scholar]
  36. Bayati, M. AI in Consulting. Bachelor’s Thesis, Degree Programme in Business Information Technology, Haaga-Helia University of Applied Sciences, Helsinki, Finland, 20 November 2019. Available online: https://www.theseus.fi/bitstream/handle/10024/266319/MutasimBayati_FinalThesis1.pdf?sequence=2&isAllowed=y (accessed on 9 February 2023).
  37. Siau, K.; Wang, W. Artificial Intelligence (AI) Ethics: Ethics of AI and Ethical AI. J. Database Manag. 2020, 31, 74–87. [Google Scholar] [CrossRef]
  38. Chui, M.; Manyika, J.; Miremadi, M. Where Machines Could Replace Humans—And Where They Can’t (yet). McKinsey Q. 2016. Available online: https://www.mckinsey.com/capabilities/mckinsey-digital/our-insights/where-machines-could-replace-humans-and-where-they-cant-yet (accessed on 29 January 2023).
  39. Whitman, C.; Sobczak, M. AI: Overrated or the Future of Accounting. p. 45. Available online: https://docplayer.net/135879214-Ai-overrated-or-the-future-of-accounting.html (accessed on 15 January 2023).
  40. Abdullah, R.; Fakieh, B. Health care employees’ perceptions of the use of artificial intelligence applications: Survey study. J. Med. Internet Res. 2020, 22, e17620. [Google Scholar] [CrossRef] [PubMed]
  41. Alexandre, C.; Blanckaert, L. The Influence of Artificial Intelligence on The Consulting Industry. Master’s Thesis, Université Catholique de Louvain, Ottignies-Louvain-la-Neuve, Belgium, 2020. [Google Scholar]
  42. Stancu, M.S.; Duțescu, A. The impact of Artificial Intelligence on the accounting profession, a literature’s assessment. Sciendo 2021, 15, 749–758. [Google Scholar] [CrossRef]
  43. Gaafar, A.S.M.; Allah, H. Artificial Intelligence in Egyptian Tourism Companies: Implementation and Perception. JAAUTH 2020, 18, 66–78. [Google Scholar] [CrossRef]
  44. Lestari, L.; Djastuti, I. Implementation of smart technology, artificial intelligence, robotics, and algorithms (STARA): A threat or opportunity for workers’ future. Rev. Manag. Entrep. 2020, 4, 149–166. [Google Scholar] [CrossRef]
  45. Zande, D. An Exploratory Case Study on the Perception towards Robotic Process Automation Among Employees. Master’s Thesis, University of Amsterdam, Amsterdam, The Netherlands, 22 June 2018. [Google Scholar]
  46. Choi, Y. A study of employee acceptance of artificial intelligence technology. Eur. J. Manag. Bus. Econ. 2021, 30, 318–330. [Google Scholar] [CrossRef]
  47. Streib, F.E. From the Digital Data Revolution toward a Digital Society: Pervasiveness of Artificial Intelligence. Mach. Learn. Knlowl. Extr. 2021, 3, 284–298. [Google Scholar] [CrossRef]
  48. Hughes, C.; Robert, L.; Frady, K.; Arroyos, A. Artificial Intelligence, Employee Engagement, Fairness, and Job Outcomes. In Managing Technology and Middle- and Low-Skilled Employees (The Changing Context of Managing People); Emerald Publishing Limited: Bingley, England, 2019; pp. 61–68. [Google Scholar]
  49. Ardon, O.; Schmidt, R.L. Clinical Laboratory Employees’ Attitudes Toward Artificial Intelligence. Lab. Med. 2020, 51, 649–654. [Google Scholar] [CrossRef]
  50. Stahl, B.C. Ethical Issues of AI. In Artificial Intelligence for a Better Future; Schroeder, D., Iatridis, K., Eds.; Springer: Berlin/Heidelberg, Germany, 2021; pp. 35–53. [Google Scholar]
  51. Chiu, Y.T.; Zhu, Y.Q.; Corbett, J. In the hearts and minds of employees: A model of pre-adoptive appraisal toward artificial intelligence in organizations. Int. J. Inf. Manag. 2021, 60, 102379. [Google Scholar] [CrossRef]
  52. Bagozzi, R.P.; Yi, Y. Specification, evaluation, and interpretation of structural equation models. J. Acad. Mark. Sci. 2012, 40, 8–34. [Google Scholar] [CrossRef]
  53. Hair, J.F.; Hult, G.T.M.; Ringle, C.M.; Sarstedt, M.; Danks, N.P.; Ray, S. An Introduction to Structural Equation Modeling. In Partial Least Squares Structural Equation Modeling (PLS-SEM) Using R.; Springer: Cham, Switzerland, 2021; pp. 1–29. [Google Scholar]
  54. Westland, J.C. Structural equation models. In Studies in Systems, Decisions and Control; Kacprzyk, P., Ed.; Springer: Berlin/Heidelberg, Germany, 2015; Volume 22, pp. 1–7. [Google Scholar]
  55. Jeon, J. The Strengths and Limitations of the Statistical Modeling of Complex Social Phenomenon: Focusing on SEM, Path Analysis, or Multiple Regression Models. Int. J. Econ. Manag. Eng. 2015, 9, 1634–1642. [Google Scholar]
  56. Lapinska, J.; Escher, I.; Górka, J.; Sudolska, A.; Brzustewicz, P. Employees’ Trust in Artificial Intelligence in Companies: The Case of Energy and Chemical Industries in Poland. Energies 2021, 14, 1942. [Google Scholar] [CrossRef]
  57. Vărzaru, A.A. Assessing the Impact of AI Solutions’ Ethical Issues on Performance in Managerial Accounting. Electronics 2022, 11, 2221. [Google Scholar] [CrossRef]
  58. Araujo, T.; Helberger, N.; Kruikemeier, S.; Vreese, C.H. In AI we trust? Perceptions about automated decision-making by artificial intelligence. AI Soc. 2020, 35, 611–623. [Google Scholar] [CrossRef]
  59. Reddy, S.; Reinartz, W. Digital Transformation and Value Creation: Sea Change Ahead. Value Digit. Era 2017, 9, 10–17. [Google Scholar] [CrossRef] [Green Version]
  60. Lichtenthaler, U. Extremes of acceptance: Employee attitudes toward artificial intelligence. J. Bus. Strategy 2020, 41, 39–45. [Google Scholar] [CrossRef]
  61. Kim, P.T.; Bodie, M.T. Artificial Intelligence and the Challenges of Workplace Discrimination and Privacy. ABA J. Labor Employ. Law 2021, 35, 289–315. Available online: https://scholarship.law.slu.edu/faculty/628/ (accessed on 8 February 2023).
  62. Acikgoz, Y.; Davison, K.H.; Compagnone, M.; Laske, M. Justice perceptions of artificial intelligence in selection. Int. J. Select. Assess. 2020, 28, 399–416. [Google Scholar] [CrossRef]
  63. Zhu, Y.Q.; Corbett, J.; Chiu, Y.T. Understanding employees’ responses to artificial intelligence. Organ. Dyn. 2021, 50, 100786. [Google Scholar] [CrossRef]
  64. Ourdedine, K. General perception of Artificial Intelligence and impacts on the financial sector in Finland. Bachelor’s Thesis, Haaga-Helia University of Applied Sciences, Helsinki, Finland, 2019. [Google Scholar]
  65. Phuoc, N.V. The Critical Factors Impacting Artificial Intelligence Applications Adoption in Vietnam: A Structural Equation Modeling Analysis. Economies 2022, 10, 129. [Google Scholar] [CrossRef]
  66. Alsheiabni, S.; Cheung, Y.; Messom, C. Factors Inhibiting the Adoption of Artificial Intelligence at organizational-level: A Preliminary Investigation. In Proceedings of the Twenty-fifth Americas Conference on Information Systems, Cancun, Mexico, 15–17 August 2019. [Google Scholar]
  67. Pan, Y.; Froese, F.; Liu, N.; Hu, Y.; Ye, M. The adoption of artificial intelligence in employee recruitment: The influence of contextual factors. Int. J. Hum. Res. Manag. 2022, 33, 1125–1147. [Google Scholar] [CrossRef]
  68. Kambur, E.; Akar, C. Human resource developments with the touch of artificial intelligence: A scale development study. Int. J. Manpow. 2022, 43, 168–205. [Google Scholar] [CrossRef]
  69. Sestino, A.; Mauro, A.D. Leveraging Artificial Intelligence in Business: Implications, Applications and Methods. Technol. Anal. Strateg. Manag. 2022, 34, 16–29. [Google Scholar] [CrossRef]
Figure 1. Configuration of the structural equation model (SEM). Source: Authors’ contribution in Stata 13.0.
Figure 1. Configuration of the structural equation model (SEM). Source: Authors’ contribution in Stata 13.0.
Electronics 12 01462 g001
Figure 2. Results of the structural equation model (SEM). Source: Authors’ contribution in Stata 13.0.
Figure 2. Results of the structural equation model (SEM). Source: Authors’ contribution in Stata 13.0.
Electronics 12 01462 g002
Table 1. Risks of artificial intelligence.
Table 1. Risks of artificial intelligence.
Bias and DiscriminationData-driven technologies have the potential to duplicate, reinforce, and magnify practices of marginalization, inequality, and prejudice that are presently found in societies. Similarly, these technologies risk replicating their developers’ preconceptions and biases because many of the features, metrics, and analytical frameworks of the models that allow data mining are selected by their developers.
Denial of Individual Autonomy, Recourse, and RightsIf individuals are subject to conclusions, forecasts, or categories created by AI algorithms, circumstances might occur where such individuals cannot hold the parties responsible for the results directly accountable. AI systems can automate cognitive tasks previously only performed by responsible human representatives. Such a lack of responsibility may impair autonomy and breach the rights of those impacted in the event of harm or bad results.
Non-transparent, Unexplainable, or Unjustifiable OutcomesAI algorithms may deliver unreliable or poor-quality results if there is irresponsible data management, negligent design or programming processes, and uncertain implementation practices. These results can directly harm the well-being of individuals and the public welfare.
Invasions of PrivacyBecause AI projects are anchored in data structuring and processing, the development of AI technologies will frequently involve personal data. These data are sometimes collected and extracted without obtaining the approval of the data owner or are processed in a way that reveals personal information.
Isolation and Disintegration of Social ConnectionAI algorithms can vastly improve consumer lives and service delivery by generating unique experiences and personalizing digital services, although this ability also has potential risks. Excessive automation could decrease the need for human interactions.
Source: Adapted from the relevant literature [23].
Table 2. Variables used in analysis.
Table 2. Variables used in analysis.
Latent VariablesObserved VariablesAcronym
Professional characteristics and the field of activity (PCFA)
Domain (management or financial consulting, accounting, financial audit, or valuation)
Domain
Work experience (under 5 years, 5–10 years, or over 10 years)
Experience
Clients (big companies, small, or both categorized by the number of employees: less or over 50 employees)
Clients
Administrative tasks required by their current position (less than 25%, 25–50%, or over 50%)
Tasks
The perceived disadvantages of using AI (DSG)
Lack of human interaction caused by using AI systems (evaluated by respondents from 1 to 5: 1—no impact; 5—critical impact)
Interaction
Job losses due to automation of tasks (evaluated by respondents from 1 to 5: 1—no impact; 5—critical impact)
Jobs
High costs and significant investments for consulting firms in building and implementing AI technologies (evaluated by respondents from 1 to 5: 1—no impact; 5—critical impact)
Costs
The lack of creativity and innovation in AI programs (evaluated by respondents from 1 to 5: 1—no impact; 5—critical impact)
Creativity
The perception of the ethical challenges of using AI (ECLG)
Discrimination in AI decision-making processes (evaluated by respondents from 1 to 5: 1—no impact; 5—critical impact)
Discrimination
The responsibility of the AI systems regarding the rights and the denial of individual autonomy (evaluated by respondents from 1 to 5: 1—no impact; 5—critical impact)
Responsibility
Privacy invasions (evaluated by respondents from 1 to 5: 1—no impact; 5—critical impact)
Privacy
Unreliable and inexplicable results (evaluated by respondents from 1 to 5: 1—no impact; 5—critical impact)
Results
Isolation and disintegration of social connection (evaluated by respondents from 1 to 5: 1—no impact; 5—critical impact)
Connection
Future dimensions of AI implementation in business consulting (AIP)
The consultants’ willingness to transfer a part of their tasks to automatic intelligent systems (operational dimension) (yes/no question)
Transfer
Their perspective on the current necessity of investments in such programs (investment dimension) (yes/no question)
Investment
Their perception of the competitive advantage gained by consulting firms that implement AI in their procedures (strategic dimensions based on hard capabilities) (yes/no question)
Competitive
Their perception of the competitive edge in recruitment processes obtained by candidates that are experienced in working with AI systems (strategic dimensions based on soft capabilities) (yes/no question)
Recruitment
Source: Own contribution.
Table 3. Results of research hypothesis testing.
Table 3. Results of research hypothesis testing.
HypothesisResultsMain Findings
H1. Professional characteristics and the field of activity determine the perceived disadvantages of using AI.Partially validated (not statistically significant)Domain and professional characteristics partially influence the perceived disadvantages of AI.
H2. Professional characteristics and the field of activity determine the future perspective of AI implementation.Partially validated (not statistically significant)Domain and professional characteristics partially influence future perspective of AI implementation in business consulting sector.
H3. The perceived disadvantages of using AI positively influences the perception of the ethical challenges of AI implementation.Validated (p < 0.001)Perceived disadvantages of AI positively influence the ethical challenges and risks: the greater the perceived disadvantages are, the greater the ethical risks and challenges become.
H4. The perception of the ethical challenges of using AI negatively influences the future perspective of AI implementation.Validated (p < 0.05)The perceived ethical challenges and risks of AI use negatively influence the willingness of AI application in professional activity: the greater the perceived ethical risks are, the lower the intention of AI implementation becomes.
Source: Authors’ contribution.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gînguță, A.; Ștefea, P.; Noja, G.G.; Munteanu, V.P. Ethical Impacts, Risks and Challenges of Artificial Intelligence Technologies in Business Consulting: A New Modelling Approach Based on Structural Equations. Electronics 2023, 12, 1462. https://doi.org/10.3390/electronics12061462

AMA Style

Gînguță A, Ștefea P, Noja GG, Munteanu VP. Ethical Impacts, Risks and Challenges of Artificial Intelligence Technologies in Business Consulting: A New Modelling Approach Based on Structural Equations. Electronics. 2023; 12(6):1462. https://doi.org/10.3390/electronics12061462

Chicago/Turabian Style

Gînguță, Andrea, Petru Ștefea, Grațiela Georgiana Noja, and Valentin Partenie Munteanu. 2023. "Ethical Impacts, Risks and Challenges of Artificial Intelligence Technologies in Business Consulting: A New Modelling Approach Based on Structural Equations" Electronics 12, no. 6: 1462. https://doi.org/10.3390/electronics12061462

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop