Abstract
Research in artificial intelligence (AI) has started in the twentieth century but it was not until 2012 that modern models of artificial neural networks aided the machine learning process considerably so that in the past ten years, both computer vision as well as natural language processing have become increasingly better. AI developments have accelerated rapidly, leaving open questions about the potential benefits and risks of these dynamics and how the latter might be managed. This paper discusses three major risks, all lying in the domain of AI safety engineering: the problem of AI alignment, the problem of AI abuse, and the problem of information control. The discussion goes through a short history of AI development, briefly touching on the benefits and risks, and eventually making the case that the risks might potentially be mitigated through strong collaborations and awareness concerning trustworthy AI. Implications for the (digital) humanities are discussed.
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
1 Introduction
It is probably safe to say that “artificial intelligence” (in short, AI) has become not only a buzzword, but also one of the most transformative technologies of our time. Together with the current trend of the digitalization of societies, it is bound to have a lasting impact on people’s lives as well as the cultures around them. Interesting models in AI are found in the domain of machine learning that are based on artificial neural networks to perform deep learning trainings and to create fresh predictions upon what the computer has “learned” from a wealth of data [77]. This makes the models potentially more powerful because they can be used for tasks that they have not been explicitly programmed for, although they have been trained to solve a specific task and broad applicability is still in its infancy. Newer models try to incorporate a form of multimodality, meaning that many different input sources are used to generate stronger potency of the algorithms for a broad applicability [2]. However, artificial neural models work with hidden layers of nodes that are in a sense opaque so that not even programmers know exactly what calculations have been performed to obtain the respective results. The main issue with this manifests itself in the problem of “AI alignment” as a concern of “AI safety engineering” [12, 17, 44, 54]. More specifically, it is known as the problem of “instrumental convergence” where a model sets itself a new instrumental goal so that it can better fulfill its original instrumental goal [14, 51]. The problem is that on the one hand, the model becomes nontransparent, and on the other hand, it may have unforeseen consequences. As such, it could eventually harm human agents and thus become misaligned with its human creators’ intentions. An example may help: an AI tasked to eliminate cancer in humans may draw the conclusion that eliminating people would be the most effective way to do so. Hence, eliminating humans would be the new goal to aid in the achievement of the original goal, which was to eliminate cancer in humans. Although such examples may sound a little too much like science fiction, the conceptual problems behind them are real and serious based on the notion that AI-agents become increasingly more self-governed and lack of explainability in their decision formation.
There are many teams working on AI safety and the large institutions in the field appear to implement their considerations responsibly. However, the speed of AI development has increased considerably not only in the past years but also in the last months and weeks. With more than just one key player on the market, there is a competitive landscape for creating ever newer and better models, which poses unique risks and benefits for society at large. Given the tremendous impact of potential machine learning issues, these risks ought to be taken seriously and it should be discussed how they could be potentially mitigated (cf. [4]).
The next chapter consists of a brief overview in the history of machine learning, predominantly focusing on illustrating how fast the recent developments have become. Based on this “economy of AI development”, some of the associated societal risks and benefits are being discussed, followed by suggestions of how these risks could be addressed.
2 The increasing speed of digital innovation in AI models
Interestingly, the idea of an “artificial intelligence” stemming from machines is not particularly new, but for a long time the approach was very different from today. Also, the advent in machine learning based on artificial neural networks was already seen in the twentieth century, although due to a lack of data and computing power, it had not grown popular until the second decade of the twenty-first century. The following discussion highlights that there was an increasing speed of digital innovations in AI models.
1940s | A necessary but not sufficient condition was the invention of the digital computer, which first saw the light of day in the 1940s in the form of the Zuze Z3 and the Atanasoff-Berry computers |
1950s | This constituted the emergence of a new discipline now known as computer science, and about a decade later, the famous mathematician Alan Turing [119] came up with what we today call the Turing-Test. The test basically tries to answer the question of when we might want to call a computer “intelligent”. Turing’s answer was that if we could not distinguish a computer’s answers from a human one, then surely, we could call this machinery intelligent. A few years later, in the mid-1950s, John McCarthy was the first person to name this an “artificial intelligence”. Needless to say, the term has all the right connotations and the potential to ignite our imagination for humans to still be intrigued by it [74, 93] Frank Rosenblatt may arguably be called the grandfather of modern AI research due to his work on the Perceptron in 1957, which appeared to be the first attempt to create an artificial neural network loosely inspired by the adaptive learning capabilities of the human brain. Although his endeavors created a considerable social hype, he did not get very far with his project because the technical limitations of his time did not allow him to move beyond one layer of hidden nodes, or, “neurons”. For this reason, the idea of artificial neural networks was abandoned rather quickly. In 1959, Arthur Samuel coined the term “machine learning” referring to a system that could play chess better than its human inventors |
1960s | In the 60s, new electronic systems were used to solve mathematical problems in the real world and applied to computers in commercial factories. A decisive moment was when Joseph Weizenbaum programmed ELIZA in 1965, which was basically the first chatbot. Now for the first time, the public was able to “speak” with a computer. Although Weizenbaum wanted to demonstrate that such interactions were extremely superficial, he later found that many people started to anthropomorphize ELIZA. Hence, already in these days, people started to theorize that one day computers might be able to think for themselves [53]. Three years later, a movie entitled 2001: A space Odyssey created a story with a machine that became sentient and eventually killed the space crew |
1970s/1980s | In the 70s and 80s, a lot of effort was made to translate the electronic capabilities onto robotics. Geoffrey Hinton—sometimes referred to as the father or the godfather of deep learning—has taken up Rosenblatt’s idea of the neural network once again and introduced more layers to it, effectively creating a multilayer perceptron (cf. [56]). This was now possible due to better computers and a revolutionary method known as backpropagation [104]. The result was that the program was able to make accurate predictions about data it has never seen before |
1980s/1990s | In the late 80s, initiatives incorporated neural networks for working on self-driving cars and in the 90s, Yann Lecun (who is also often counted amongst the AI pioneers and was working alongside Hinton) created a program that recognized hand-written digits. However, for at least another decade the field was stifled by the lack of computing power and the lack of huge data sets. At this time, skeptical voices claiming that AI was not as promising as once thought led to a decline in funding, eliciting what is sometimes referred to as an a “AI winter” [53, 134] |
2012 | The birth of the modern AI movement can be traced back to September 2012. This is when Hinton and his team were now operating in a world with computers that were fast enough and had enough data due to the internet so that they were able to create AlexNet, an image recognition system that could solve the ImageNet challenge (a large collection of images that had to be labelled independently by the computer) better than any previous technology. The impact this finding had can hardly be overstated. The results were highly promising, which led to a large-scale global adoption of artificial neural networks and subsequently many innovations in the domain emerged [64]. Today, neural network models have become commonplace: Tesla uses it for self-driving cars, Google uses it to learn about our search parameters, Netflix and YouTube use it to estimate which movies we like, Amazon employs it to find what we are likely to buy, Uber has it to allocate the driver as well as to calculate the pricing and the roads, and every year, new start-ups come on to the scene employing new AI innovations. Many apps like TikTok, Instagram or Tinder are completely AI-driven, and the same is becoming more mainstream in the physical space |
2015/2016 | In 2015, DeepMind’s Alpha Go has won against the European champion Fan Hiu and [47] it defeated Lee Sedol, the world champion, in a very complex game called “Go”. The program learned the game all by itself with no supervision |
2017/2018 | These exponential AI improvements have become palpably visible especially in the past 6 years. In August 2017, Google introduced the so-called transformer model, which has been foundational for LLMs (large language models) to emerge [124]. Based thereupon, in June 2018 the generative pre-trained model GPT-1 was released by OpenAI [97] |
2019 | Not even a year after this, OpenAI leveled up their LLM in February 2019 and released the largest model at the time with 1.5 billion parameters (Bp) called GPT-2 [98]. Then, in October 2019, Google released BERT, which significantly improved search queries [87] |
2020 | Google AI made a huge innovation public in January 2020 with Meena, a chatbot that is conversant in most everyday topics [1]. Only three months later, Facebook’s Meta AI published BlenderBot 1.0, an even stronger chatbot model, which was made available completely open-source [102]. A month after this, OpenAI published GPT-3, today one of the most powerful language models, with a staggering 175 Bp as well as a few- and zero-shot learning approach [16]. In September 2020, The Guardian in collaboration with OpenAI published an article that caught the public’s attention because it was fully written by the AI ([48], p. 3)Footnote 1 |
2021 | In the following two years, the exponential developments have increased even more, due to increasing computing power, big data, and model innovations [38, 134]. In January 2021, EleutherAI issued The Pile v.1, an 800 GB dataset of diverse text for language modelling, aiding future research in the domain [45]. In March, BAAI published Wudao 1.0, which is a “ super large-scale Chinese corpora for pre-training language models” [136]. In June, EleutherAI introduced a new generative pre-trained transformer called GPT-J-6B, which was developed in Japan and had 6 Bp [130]. A few days later, Google used the improvements in LLMs to publish a new chatbot model for dialog applications known as LaMDA (for a revised version, see [117]. Around this time, the Chinese company BAAI updated their product with Wudao 2.0. [141]. Briefly after, the Chinese institution Alibaba Dharma Academy published the M6, an innovative model due to its usage of multimodal databases including texts and images [72]. Just a month later, Facebook’s Meta introduced their improved chatbot model, the BlenderBot 2.0 consisting of 9.4 Bp [65]. Likewise, DeepMind released an open-source license for AlphaFold, an AI that can predict protein structures better than humans can, which is immensely valuable in the fields of biology and medicine. In August, AI21 published the largest LLM with 178 Bp called Jurassic-1 [70], Baidu introduced Plato-XL with 11 Bp with the clear intent of anthropic alignment [7], and Allen AI/AI2 released an 11 Bp QandA-model [111]. In October, NVIDIA and Microsoft hit the new record with Megatron-Turing NLG using 530 Bp (cf. [108], followed by Yuan 1.0 with 245 Bp by Inspur AI from China [135]. In November, Alibaba Dharma Academy revised their M6 model [72] and Coteries in France issued Cedille, a 6 Bp model based on GPT-J-6B [80]. In December 2021, a new player known as Anthropic published their 52 Bp model (cf. [7], Google Inc. and Google AI broke the record with GLaM, a model with 1.1 terra parameters [25] as well as with Gopher, which was noteworthy because it incorporated 280 Bp with novel multimodal dataset [99]. The year went to an end with a Christmas gift from Meta AI using a 13 Bp model named Fairseq-13B [6] |
2022 | In 2022, OpenAI implemented new data resulting in a new variant of GPT-3 known as text-davinci-002 [91], followed by GPT-NeoX-20B from EleutherAI in February [11, 68]. In March 2022, Google’s DeepMind released Chinchilla with 70 Bp [57], BigScience published tr11-176-ml (only available on the github repository). April brought along PaLM with 540 Bp by Google [83], CodeGen using 16 Bp by Salesforce, VLM-4 by Salesforce [88], the 200 Bp Luminous by Aleph Alpha [52], the 13 Bp mGPT from Sber [107], and the 10 Bp Noor from TII [114]. The most groundbreaking model in April was called Flamingo produced by Google’s DeepMind because it combined a 70 Bp language model with a 10 Mb image model [3]. In May, Meta AI released OPT-175B with an open-sourced code ([137, 138]. And just in the past few days at the time of this writing, Google AI published LaMDA 2 [117], DeepMind released an intriguing model called Gato with 1.18 Bp [100], and Google Research published UL20B [113] At the time of this publication, the last few weeks showed groundbreaking innovations in multimodality by combining natural language processing (NLP) with computer vision, the most prominent models are Dall-E 2 by OpenAI [105], Flamingo by DeepMind [3], and a few days ago, Imagen was introduced by Google [105]. All these models are incredibly powerful in using text inputs to generate images or taking images and converting them to natural text |
2022/2023 | Around the switch from the year 2022 to 2023, an interesting business move was made that came along with a considerable social impact. So far, most of these developments had occurred behind an ‘expert veil’ since the general public neither was up to date on the most prominent models (except from anecdotes and scattered media reports) nor was there large-scale interest in them since access was mostly reserved for professionals working in the field as well as some selected beta-testers. In November 2022, [92] decided to go public with a version of GPT-3.5 in a chatbot graphical user interface (GUI) and provide free access to the world. Within days and weeks, global news agencies reported on the introduction of the application dubbed ChatGPT whereas people from all industries started to adopt the technology for their business ideas, writing papers and articles, generating slogans, and helping out with tasks requiring natural language applications [42]. Within just a short period of time, the global community was faced with the questions of how to best deal with such AI models and how to regulate them [41, 50, 96, 140]. At the beginning of 2023, Microsoft [78] as a prominent investor in OpenAI has added to the momentum by announcing that they planned to incorporate GPT into their search engine Bing. Fathoming their business model threatened, at the same time Google publicly declared that they intended to stay ahead of the game and thus to also implement their NLP model called Bard for their search engine [94]. This provoked some popular commentators on the web to proclaim that the “AI wars” have begun [18] |
This is supposed to show just how accelerating the speed of AI development has become (an up to date timeline of AI and language model developments can be found here: [116].
3 The rapidly changing economy of AI development
On his YouTube channel, Alan Thompson, a leading researcher in the field, stated at the end of May 2022:
I documented just how quickly this stuff is changing. In March, April and May 2022, we were generating something like a new model every three to four days coming out from places like DeepMind, Google Research, OpenAI, EleutherAI, and the list goes on. It’s almost a full-time job to keep up with it [115].
With this rapid acceleration in the development of the AI economy, there are some benefits and risks emerging. The potential risks are of notable relevance because they might have a considerable impact on human societies, which is why it is important that they are being discussed. In the current paper, the quintessence of the AI economy is conceived in the fact that there is a competitive market of developers that have new ideas for models, data usage and different access to computing power to create new AI systems. Eventually, there appears to be a uniting goal, which is to be the first to arrive at an Artificial General Intelligence (AGI). This is, metaphorically speaking, a race to the moon.
3.1 Benefits of AI competition
In digital business models, the idea of “the winner takes it all” appears to have a lot of merit [66]. The reason for this is that inventors of digital platforms and technologies control not only the access to their know-how and the new possibilities, but they also gain a competitive advantage where customers are willing to pay more for a service than will be the case later on, and as soon as they are connected to the innovator’s infrastructure, it becomes increasingly more difficult to switch partners [125]. As such, the key developers have an inherent incentive to invest heavily in these new technologies and to be faster than their competitors [106, 133].
The benefits in these dynamics largely lie in the associated market dynamics. For a long time, the popularity of machine learning research was confined to a small population of experts. The limitations in data availability as well as computing power did not seem to make the field very attractive and thus, there were not as many players on the market as today [39]. This had three consequences [82]: (i) funding was capped, (ii) human resources were limited, and (iii) there was less potential to share knowledge with other researchers and hence make faster progress together. Economically, this last point has a two-sided effect. On the one hand, a larger market may motivate the agents to collaborate and to share their knowledge to better keep up with the fast developments. On the other hand, each protagonist has the business incentive to be faster than the others so that one can reap the benefits from the major AI breakthroughs. This leads to an ever-increasing acceleration of AI development as long as the players on the market believe that they are close to a profitable turnaround [34, 43].
The main benefit may be the innovations in new machine learning models. One could argue that a fast improvement in these systems eventually profits the consumers and therewith human societies at large since AI solutions can be applied in virtually every setting we dare to imagine [32]. Hence, if there were no risks to them, it would be admirable to get ever-increasingly fast developments until a singularity is achieved (which is the idea that AI systems would surpass human intelligence).
3.2 Risks of AI competition
Of course, market and business principles do not explain all the developments in the field. There are major ethical concerns governing some of the decisions of the leading players. OpenAI, one of the major institutions in the field and the authors of GPT-3, chatGPT and Dall-E 2, is deliberately set up as a hybrid organization between a business company and an NPO with philanthropic goals. They portray a sense of urgency, which becomes apparent in their mission statement where they claim that the top priority of AI development must lie in safety. The team at OpenAI vows that if any of their rivals comes close to achieving a true AGI, they would cease their activities and concentrate on collaborating with this competitor to see the project come to fruition [90]—although it is a philosophical question when (if ever) one would be inclined to grant a system the label of an AGI.
There are three central concerns that should be acknowledged in developing potent AI systems, all of them being manifestations of a discipline known as “AI safety engineering” (for a glimpse into this discipline, see [12, 17, 44, 54]. The first deals with the problem of AI misalignment, the second with the problem of human abuse, and the third with the problem of information control (cf. [127]).
AI alignment is the study of finding ways to make sure that an AI’s instrumental goals do not conflict with our human terminal goals. The previous example of eradicating cancer by getting rid of humans altogether would be a manifestation of clear misalignment where the human goals are not adequately reflected in the AI’s objectives. Since in machine learning, the system can “learn” by itself, it also has the possibility to implicitly set itself new goals to fulfill its programming effectively. Since it is complicated to peek into a system’s hidden layers, it is also not obvious whether an AI has established such a latent goal that might be at odds with human objectives. These things need to be weighed very carefully if a general AI is to be released “into the wild” [19, 139].
Even if there is perfect alignment with the human intentions, the intentions of individuals may not always be for the greater good. This might occur either through ignorance or through human maliciousness. A person could task an AI to do harmful things (either willingly or unwillingly) and hence create damage in the real world harming real people. This is not a problem if researchers experiment with machine learning algorithms in their labs. But as soon as people could access these technologies more broadly, they also would have the potential to abuse them if they wanted to. And just with anything else, there will surely be some people who will want to use these technologies with harmful intentions [28, 75, 127].
The third problem has to do with explainability and information control, meaning that the AI has the capacity to retrieve data but also to create brand new information, either in text but also in picture or video form. To a certain degree, this is already possible with OpenAI’s Dall-E 2, DeepMind’s Flamingo and Google’s Imagine. Soon, it will be impossible to discern between fact and fiction in the digital sphere because everybody would have access to powerful data generators based on AI. This is at least part of the reason why many players have refused to give the public access to these applications at the beginning (cf. [21, 84]).
These topics in AI safety engineering are highly relevant, but chances are high that in an increasingly accelerating AI economy, the speed might lead to an underrepresentation of these risks. This would lead to AI being used by the public or by controversial protagonists prematurely, before either the technology or society would be ready [5, 9, 46, 62, 63, 126]. There are two ways how this could come about: first, the incentives of “going live” might start to overrule the AI safety concerns of the current players, or second, the speed increases the prospects for newcomers so that new agents enter the market that might not share the same concerns because they could be driven by short-term incentives.
3.3 Managing the risks
These rapid AI developments harbor their own economic dynamics, which raises the question of how the associated risks can be managed and mitigated. The risks deal with an underrepresentation of AI safety concerns [127]. The main question, then, is how to ensure that the focus on trustworthiness remains a pivotal concern, both in the books of large incumbents as well as when it comes to newcomers entering the field.
Via EU funds from the Horizon 2020 program, the European Union has attempted to foster developments in the direction of socially acceptable machine learning tools. There are four European networks aspiring research excellence in AI, which are called ELISE, HumanE-AI, AI4Media, and TAILOR. The latter was deliberately launched to work on the topic of trustworthy AI and to establish a network of research centers that can act as expert hubs in the domain. In an interview with Inria, Marc Schoenauer (project-team leader of TAU, an initiative involved in TAILOR), declared: “If systems lack transparency, are biased, are easily affected by noise or attacks, or if their results cannot be explained, then society will never accept AI.”([60], para. 5).
There are ethics guidelines presented by the European Commission [36] stating that trustworthy AI should be:
-
(1)
Lawful—respecting all applicable laws and regulations.
-
(2)
Ethical—respecting ethical principles and values.
-
(3)
Robust—both from a technical perspective while taking its social environment into account.
These points were further specified by seven key requirements for an AI system to be deemed trustworthy. They are the following:
-
Human agency and oversight:
AI systems are supposed to help humans to make informed decisions and at any time, proper oversight mechanisms must be ensured, such as human-in-the-loop, human-on-the-loop, or human-in-command approaches.
-
Technical robustness and safety:
AI systems should be secure, accurate, reliable, and reproducible. Additionally, there must be a fallback plan in case something goes wrong.
-
Privacy and data governance:
The quality and integrity of data should be respected, making sure that access is restricted to the respective agents involved.
-
Explainability:
The decisions of AI systems should be explainable in a manner adapted to the concerned stakeholders. People always have to be aware that they are interacting with an AI and they should be informed about the model’s capabilities and limitations.
-
Diversity and fairness:
Unfair biases and discrimination must be avoided. AI systems should be accessible to all, regardless of nationality or other demographic variables.
-
Societal and environmental wellbeing:
AI systems should come to the advantage of all people, both in the present and the future. Environmental and social impacts should be carefully considered.
-
Accountability:
There should be mechanisms to guarantee responsibility and accountability of AI systems and outcomes. There ought to be ways to audit the systems, enabling assessment of the algorithms, the data, and processes in place.
Based on these principles, the TAILOR network was formed, which now includes 55 partners across Europe. Its vision is to achieve human-centered trustworthy AI and make Europe the global role-model for responsible AI. Although the aspiration is noble, considering that a lot of the innovation in the field does not only originate from Europe (cf. chapter 2), it may be unlikely that the continent itself will be “the” role-model in the field, but might still be a contributing force. The network is built around six core objectives [112] (Table 1).
The EU guidelines have been criticized as being somewhat myopic, vague, and having a lack of long-term risk considerations. Furthermore, “red lines” like autonomous lethal weapons and social scoring systems had not been given enough priority [76]. They were criticized as having no hierarchy of principles [101], being non-binding, providing little incentives to adhere to the ethical principles, and having a lack of regulatory oversight to monitor the enforcement of EU ethical guidelines [73]. Although useful for providing a starting point concerning trustworthy AI, companies and practitioners worry that the EU guidelines are not specific enough for the harmonization with company law and governance principles, as well as leaving many concrete questions unanswered [55]. A review on the timeline in the European legislation of AI concluded that it is going to be a long road for the specific formulation and implementation since there are many open questions that must be dealt with in a nuanced manner [40]. Another reason seems to be that it may be difficult for regulators to keep up with the pace of the new inventions of modern AI models that keep changing (see chapter 2).
When perceived in light of the rapid advancements in AI development and the associated economic risks in safety engineering as highlighted above, not all TAILOR objectives appear to be equally promising. The first objective is geared towards a network of research centers, which is helpful for advancing knowledge in the field, but does not necessarily mitigate the risk of the developmental speed leading to a loss of vision on safety engineering for certain companies. The reason is that these research centers and the individual AI organizations may not necessarily overlap or collaborate. Similar thoughts can be applied to the second objective. It sets out to define and maintain a strategic roadmap, which is highly relevant for the excellence centers in the TAILOR network, but it is not binding for all external institutions. The third objective wants to create the capacity and critical mass to develop the scientific foundations for trustworthy AI. There are two problems with this: (i) the most promising resources and technologies for generating large statistical models with more than one billion parameters are mostly in the hands of established companies in the industry whereby academic institutions neither have the funds to currently calculate the models nor to keep up with the speed; (ii) and even if TAILOR could generate a trustworthy AI, this would do nothing to impede other industry players that do not abide by the same ethical guidelines. The fifth objective runs into the same problem of that progress in the state-of-the-art foundations of trustworthy AI is only relevant for the players that already subscribe to these ethical tenets. Much more promising are the fourth and sixth objectives. Objective four intends to build strong collaborations between academic, industrial, governmental, and community stakeholders. These collaborations can act as platforms or gatekeepers providing access to the necessary resources, know-how and market touchpoints in the field. If the networks become global and exclusive in the sense that it is almost impossible to succeed without being a part of them, then the networks’ ethical guidelines become a mandatory part of the AI developments, regardless of the speed they take on. Comparable to this, objective six wants to increase knowledge and awareness in the foundations of trustworthy AI, which can influence stakeholders to make sure that the companies abide by established ethical guidelines.
Although the EU has invested a lot of resources striving for laws in the regulation towards trustworthy AI, it is by far not the only regulator to do so. As (Smuha [109], p. 16) stated:
Europe has not been alone in making strides towards addressing the ethical concerns raised by AI systems – and that is a good thing. […] Discussions with other countries and regions, taking place at global fora such as the G7 and the G20, and within other organisations such as the OECD, the Council of Europe and UNESCO, are essential to work towards a much-needed global governance framework for AI.
The initiatives started to become more pertinent [47] when China announced its three-year guidance for the internet and AI plan [20], whereas the US issued their national AI research and development strategy plan [86], and South Korea published the mid- to long-term master plan in preparation for the intelligent information society [47]. In 2017, Canada [22], Japan [110], Singapore [59], Finland [79], the United Arab Emirates [85], and China again [131] issued further regulatory strategic initiatives. In the consecutive years, a host of other countries from Europe, Northern America and Asia began to publish digital strategies and plans for regulations dealing with AI (for a timeline review, see [123]). International collaborations were fostered, among others, by the EU’s communiqué on AI in 2018 [35], the OECD principles on AI from 2019 [89], the EU’s white paper on AI in 2020 [37], and the UN’s strategic plan for AI and robotics for 2020–2025 [122].
A recent research paper on the principles and practices in trustworthy AI highlights that most of the developmental processes and operational advancements occur in the industry. The companies form the sector that eventually connects the users, academia, as well as the government with the market. In the industry, operation occurs along the interfaces of data management, algorithm design, development, and deployment [69]. As such, it is necessary that initiatives working on trustworthy AI form collaborative platforms with companies from the industry that are key players in the market.
3.4 Implications for the (digital) humanities
Almost all academic enterprises are affected by these developments. This is true even for disciplines where this might seem somewhat counterintuitive, like the humanities. Since the speed of innovation is permeating society unhalted, it may elevate critical thinking faculties and the (digital) humanities to a particularly valuable place in these ongoing dynamics. This last chapter poses an example of what this might mean for the humanitarian disciplines.
There are two directions where the current AI developments may become relevant for the humanities: first, the digital humanities can and should preserve a critical humanitarian glimpse upon the rapid machine learning advances that are likely to have a transformative impact on our societies. Second, the humanities by and large may need to evaluate how these evolving technologies could be implemented in their own fields of study in the most promising fashion. Both of these points shall be briefly discussed in the following.
As seen in chapter 2, the AI models are being developed at an accelerating speed. At the time of this writing, in the past few months almost every day a new machine learning model with novel features was published. Already today, custom-tailored digital catering via algorithms defines which books we read, which financial investments we make, which movies we watch, which doctors we consult, which news articles we consume, which food we eat, and even which dates we go on (i.e., through algorithmic matching via Tinder, Bumble and Co). Since machine learning models are getting larger, faster, better and more multimodal, the transformation they will have on our societies is likely to impact every aspect of human life and conversely our social interactions (cf. [128, 129]). There is the danger that technological engineers, software developers and businesspeople do not necessarily have the same goals and incentives as individual consumers, ordinary citizens, policy makers, or societies at large. As such, given the rapid advancements, there may be the risk that the needs of some stakeholders would be overheard [13, 15, 27, 31, 71, 95]. This is where digital humanists can assume a crucial role in these dynamics: with a focus not primarily on profit or technological concerns, they should research and discuss how such changes impact humans and societies, preserving a critical stance towards these developments and at the same time making constructive suggestions to help engineers who often have less time to dwell on ethical concerns, or have to meet deadlines and revenue goals. This has been described as a foundational task of the digital humanities, which would be a valuable contribution to the community [10, 61, 67]. In line with the above discussion, for this to work best it appears necessary that digital humanists stay in close connection with researchers and practitioners from the industry, so that (i) they remain in touch with the rapid changes in the field, and (ii) their critical reasoning influences the discussion as well as the decision-making exactly in the places where the AI systems are being constructed.
At the same time, due to the broad applicability and pervasiveness of AI models, they are bound to play a significant role in all academic disciplines. It is likely that no field will be omitted from these dynamics, which might be comparable to the universal introduction of computers in the 1940s. Individuals, regardless of whether this is in their private lives, in companies or in academia, have become accustomed to the fact that computers are often integrated in what humans do. In the past hundred years and more, computers have not taken away our responsibility to think, act and make decisions, but they have helped us in doing so. They are helping us with storing, processing, and retrieving information, as well as translating some of them to physical labor via robotics. However, due to the inherent limitations of classical computing systems, they still require humans to engage in a multitude of micro-decisions. If machine learning algorithms now are becoming increasingly better, they will push these boundaries further aback, allowing us to hand over more of the repetitive work to the machines. This does not imply that the work humans do becomes mindless or less credible. It only means that there needs to be a form of “AI literacy” when working with them. Neither the industry not science has become worse by enabling all workers to use computers. In fact, the opposite is the case: there are more options to solve problems, and tasks can be managed more effectively. What is necessary, though, is that people who use computers need to know what they can and cannot do with them. The argument here is that this would be no different with the introduction of AI systems. After all, modern machine learning algorithms are nothing else than statistical models based on a huge set of matrix multiplications, and there are publications highlighting the inherent limits of such systems. For example, it lies in the nature of their architecture that AI models will never be able to do things that are quite simple for humans, such as making logical inferences or entertain abstract reasoning—because all they do (and this is what they are really good at) is making extrapolations using statistical operations [29, 137, 138]. However, such claims may be debatable since they strongly depend on what we mean when making reference to terms like “abstract reasoning” or also just “reasoning” per se [26, 49].
In the humanities, deep neural networks have already been applied in a variety of tasks, i.e. Topic Modelling [8, 81, 118] or Authorship Verification [58, 132]. Three years ago, a paper called “The computational case against computational literary studies” by Nan Z. Da [23] has made some headlines by claiming that the literary studies have used neural models incorrectly. One of the points made was that this further exacerbates the replication crisis in the discipline. Kathryn Eccles ([33], p. 90) put it well when saying that “Da’s article had an immediate impact”. It provoked numerous responses with many scholars defending the approaches used in the past. A few days later, Da [24] reiterated the claim in her essay entitled The Digital Humanities Debacle, whereas Ted Underwood [121] pushed back in his defense called Dear Humanists: Fear Not the Digital Revolution. James E. Dobson [30] suggested a synthesis of the two perspectives by stating that going forward, both qualitative criticism as well as computational statistics will be valuable for further advancements in the field. Given the speed of AI development and its potential pervasiveness (see chapter 2), just like in any other discipline, it appears unrealistic that the humanities will not be working with AI models; almost like most of us today work with computers in our everyday jobs. Today, computers do not take away our responsibilities to think, act and make decisions, but they help us to be better informed and to work more efficiently. The same will arguably be the case with the general adoption of AI systems.
Humans and AI will likely work together hand-in-hand, akin to our symbiotic relationship to computers nowadays. Since machine learning algorithms are (i) only statistical models and (ii) remain opaque in their inner workings due to the multitude of interconnected nodes (or “neurons”), they can provide us with important clues but not with definite and secured knowledge. This means that humanists should not simply take an output at face value but as valuable information providing statistical evidence for certain thoughts. The same holds true in the natural sciences: engineers are training AI systems to recognize cancer cells on MRI images. This does not make the physician unnecessary, but instead shifts his or her job to becoming a controlling agent, making sure that the computer has found the right solutions (mostly checking for false positives), thereby making the whole process of diagnosing cancer much more efficient and accurate [120]. In other words, the AI provides valuable information whereas the physician checks the results and makes the necessary decisions afterwards. Another example is DeepMind’s AlphaFold, which is an AI that in many cases works better in analyzing protein folding structures than the way molecular biologists hitherto had to disentangle them. This leads to groundbreaking results in biology and medicine. However, it does not imply that the structures can be taken at face value—the results need to be carefully tested experimentally in the lab to confirm that the folding occurs the exact way that AlphaFold is suggesting. The emergent human-AI interactions produce findings that would have taken decades for biologists to discover, meaning that the whole process becomes a lot more effective [38, 103].
These ideas can be applied to the (digital) humanities: neural network models are not supposed to force a result upon a researcher. The results obtained by an artificial neural network are fuzzy, depending on the learning algorithm not always reproducible, and it is unclear what calculations were performed due to the nature of deep learning architectures. This problem is not unique to any given discipline. However, human experts can use the outputs as evidence and make sense of them by framing the results with their expertise. As such, humanity researchers can—and most likely inevitably will—work with AI systems, just as they today also work with computers.
4 Conclusion
The present paper intended to give a brief overview through the history of AI development, with a focus on demonstrating that there has been a rapid acceleration in these dynamics, especially in the past few months and weeks. Although there are certainly some benefits in these fast market developments, there are also some risks that belong to the domain of AI safety engineering. The main concerns are an improper AI alignment, the abuse of such technologies and problems with informational control (that is, discerning fact from fiction in the digital space). It is important to work on solutions to mitigate these risks on the levels of politics and the law as well as the society. Some ethics guidelines proposed by the EU commission were discussed and framed by TAILOR, an EU initiative working on trustworthy AI. The present paper discussed that some of the TAILOR objectives are more promising than others, given the risks previously discussed. The most promising objectives are the ones dealing with building strong collaborations, and the ones dealing with raising awareness in the domain of trustworthy AI. There are some implications that can be deduced for the (digital) humanities. First, it was claimed that digital humanists must preserve a critical glimpse upon these rapid changes occurring in the world of AI development, making sure that ethical issues remain prominently discussed. Second, it was held that AI systems will become commonplace in both the industry and academia, meaning that humanity experts will have to figure out how these new technologies can be adequately integrated. The suggestion was made that outputs from deep neural networks should not be taken at face value and cannot replace qualitative criticism, meaning that there should eventually be a nuanced expert-AI interaction.
Data availability statement
There were no datasets used that were specific to this publication.
Notes
This raises interesting questions to be pondered: how should one cite an article that was entirely written by an AI? Should one merely cite the name of the language model or also somehow refer to the humans who entered the prompts? And what if the prompt to create the article might have come from yet another AI?
References
Adiwardana, D.: Towards a conversational agent that can chat about…anything [google brain research report]. Google AI Blog. http://ai.googleblog.com/2020/01/towards-conversational-agent-that-can.html (2020). Accessed 26 May 2022
Afshar, M., Sharma, B., Dligach, D., Oguss, M., Brown, R., Chhabra, N., Thompson, H.M., Markossian, T., Joyce, C., Churpek, M.M., Karnik, N.S.: Development and multimodal validation of a substance misuse algorithm for referral to treatment using artificial intelligence (SMART-AI): a retrospective deep learning study. Lancet Digit. Health 4(6), e426–e435 (2022). https://doi.org/10.1016/S2589-7500(22)00041-3
Alayrac, J.-B., Donahue, J., Luc, P., Miech, A., Barr, I., Hasson, Y., Lenc, K., Mensch, A., Millican, K., Reynolds, M., Ring, R., Rutherford, E., Cabi, S., Han, T., Gong, Z., Samangooei, S., Monteiro, M., Menick, J., Borgeaud, S., et al.: Flamingo: a visual language model for few-shot learning (arXiv:2204.14198). arXiv (2022). https://doi.org/10.48550/arXiv.2204.14198
Amodei, D., Olah, C., Steinhardt, J., Christiano, P., Schulman, J., Mané, D.: Concrete problems in AI safety (arXiv:1606.06565). arXiv (2016). https://doi.org/10.48550/arXiv.1606.06565
Araujo, T., Helberger, N., Kruikemeier, S., de Vreese, C.H.: In AI we trust? Perceptions about automated decision-making by artificial intelligence. AI Soc. 35(3), 611–623 (2020). https://doi.org/10.1007/s00146-019-00931-w
Artetxe, M., Bhosale, S., Goyal, N., Mihaylov, T., Ott, M., Shleifer, S., Lin, X.V., Du, J., Iyer, S., Pasunuru, R., Anantharaman, G., Li, X., Chen, S., Akin, H., Baines, M., Martin, L., Zhou, X., Koura, P.S., O’Horo, B., et al.: Efficient large scale language modeling with mixtures of experts (arXiv:2112.10684). arXiv (2021). https://doi.org/10.48550/arXiv.2112.10684
Askell, A., Bai, Y., Chen, A., Drain, D., Ganguli, D., Henighan, T., Jones, A., Joseph, N., Mann, B., DasSarma, N., Elhage, N., Hatfield-Dodds, Z., Hernandez, D., Kernion, J., Ndousse, K., Olsson, C., Amodei, D., Brown, T., Clark, J., et al.: A general language assistant as a laboratory for alignment (arXiv:2112.00861). arXiv (2021). https://doi.org/10.48550/arXiv.2112.00861
Asmussen, C.B., Møller, C.: Smart literature review: a practical topic modelling approach to exploratory literature review. J. Big Data 6(1), 93 (2019). https://doi.org/10.1186/s40537-019-0255-7
Bartz-Beielstein, T.: Why we need an AI-resilient society (arXiv:1912.08786). arXiv (2019). https://doi.org/10.48550/arXiv.1912.08786
Berry, D.M., Fagerjord, A.: Digital Humanities: Knowledge and Critique in a Digital Age. John Wiley and Sons, Cambridge (2017).
Black, S., Biderman, S., Hallahan, E., Anthony, Q., Gao, L., Golding, L., He, H., Leahy, C., McDonell, K., Phang, J., Pieler, M., Prashanth, U.S., Purohit, S., Reynolds, L., Tow, J., Wang, B., Weinbach, S.: GPT-NeoX-20B: an open-source autoregressive language model (arXiv:2204.06745). arXiv (2022). https://doi.org/10.48550/arXiv.2204.06745
Boggust, A., Hoover, B., Satyanarayan, A., Strobelt, H.: Shared Interest: Measuring Human-AI Alignment to Identify Recurring Patterns in Model Behavior. In: CHI Conference on Human Factors in Computing Systems, pp. 1–17. https://doi.org/10.1145/3491102.3501965 (2022)
Bogina, V., Hartman, A., Kuflik, T., Shulner-Tal, A.: Educating software and AI stakeholders about algorithmic fairness, accountability, transparency and ethics. Int. J. Artif. Intell. Educ. 32(3), 808–833 (2022). https://doi.org/10.1007/s40593-021-00248-0
Bostrom, N.: The superintelligent will: motivation and instrumental rationality in advanced artificial agents. Mind. Mach. 22(2), 71–85 (2012). https://doi.org/10.1007/s11023-012-9281-3
Brennen, A.: What Do People Really Want When They Say They Want “Explainable AI?” We Asked 60 Stakeholders. In: Extended Abstracts of the 2020 CHI Conference on Human Factors in Computing Systems, pp. 1–7. https://doi.org/10.1145/3334480.3383047 (2020)
Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J.D., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., Agarwal, S., Herbert-Voss, A., Krueger, G., Henighan, T., Child, R., Ramesh, A., Ziegler, D., Wu, J., Winter, C., et al.: Language Models are Few-Shot Learners. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 1877–1901. Curran Associates, Inc. https://proceedings.neurips.cc/paper/2020/file/1457c0d6bfcb4967418bfb8ac142f64a-Paper.pdf (2020)
Butlin, P.: AI Alignment and Human Reward. In: Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society, pp. 437–445. Association for Computing Machinery. https://doi.org/10.1145/3461702.3462570 (2021)
Cai, K.: The New AI Wars Begin As Microsoft Announces ChatGPT Integration with Bing And Edge. Forbes. https://www.forbes.com/sites/kenrickcai/2023/02/07/the-new-ai-wars-begin-as-microsoft-announces-chatgpt-integration-with-bing-and-edge/ (2023). Accessed 23 Feb 2023
Carlsmith, J.: Is power-seeking AI an existential risk? (arXiv:2206.13353). arXiv (2022). https://doi.org/10.48550/arXiv.2206.13353
Chinese State Council: China Releases 3-year Guidance for Internet Plus AI Development [GOV-CN]. The People’s Republic on China. http://english.www.gov.cn/news/video/2016/05/24/content_281475356234629.htm (2016). Accessed 26 Jan 2023
Chromik, M., Eiband, M., Buchner, F., Krüger, A., Butz, A.: I Think I Get Your Point, AI! The Illusion of Explanatory Depth in Explainable AI. In: 26th International Conference on Intelligent User Interfaces, pp. 307–317. https://doi.org/10.1145/3397481.3450644 (2021)
CIFAR: The Pan-Canadian AI Strategy [Official Hoempage]. Canadian AI Research Center. https://cifar.ca/ai/ (2017). Accessed 26 Jan 2023
Da, N.Z.: The computational case against computational literary studies. Crit. Inq. 45(3), 601–639 (2019). https://doi.org/10.1086/702594
Da, N.Z.: The Digital Humanities Debacle [The Review]. The Chronicle of Higher Education. https://www.chronicle.com/article/the-digital-humanities-debacle/ (2019b). Accessed 30 June 2022
Dai, A.M., Du, N.: More Efficient In-Context Learning with GLaM [Research Report]. Google AI Blog. http://ai.googleblog.com/2021/12/more-efficient-in-context-learning-with.html (2021). Accessed 26 May 2022
Davis, E., Marcus, G.: Commonsense reasoning and commonsense knowledge in artificial intelligence. Commun. ACM 58(9), 92–103 (2015). https://doi.org/10.1145/2701413
Deshpande, A., Sharp, H.: Responsible AI Systems: Who are the Stakeholders? In: Proceedings of the 2022 AAAI/ACM Conference on AI, Ethics, and Society, pp. 227–236. https://doi.org/10.1145/3514094.3534187 (2022)
Diakopoulos, N., Johnson, D.: Anticipating and addressing the ethical implications of deepfakes in the context of elections. New Media Soc. 23(7), 2072–2098 (2021). https://doi.org/10.1177/1461444820925811
Dickson, B.: Large Language Models have a Reasoning Problem [Blog Forum]. TechTalks. https://bdtechtalks.com/2022/06/27/large-language-models-logical-reasoning/ (2022). Accessed 30 June 2022
Dobson, J.E.: Critical Digital Humanities: The Search for a Methodology. University of Illinois Press (2019)
Du, S., Xie, C.: Paradoxes of artificial intelligence in consumer markets: ethical challenges and opportunities. J. Bus. Res. 129, 961–974 (2021). https://doi.org/10.1016/j.jbusres.2020.08.024
Duan, Y., Edwards, J.S., Dwivedi, Y.K.: Artificial intelligence for decision making in the era of Big Data–evolution, challenges and research agenda. Int. J. Inf. Manag. 48, 63–71 (2019). https://doi.org/10.1016/j.ijinfomgt.2019.01.021
Eccles, K.: Digital humanities. Year’s Work Crit. Cult. Theory 28(1), 86–101 (2020). https://doi.org/10.1093/ywcct/mbaa014
Elliott, A.: The Culture of AI: Everyday Life and the Digital Revolution. Routledge. https://doi.org/10.4324/9781315387185 (2018)
European Commission: Artificial Intelligence for Europe (Communication from the Commission SWD-137 final; pp. 1–20). European Union. https://eur-lex.europa.eu/legal-content/EN/TXT/?uri=COM%3A2018%3A237%3AFIN#document1 (2018). Accessed 26 Jan 2023
European Commission: Ethics guidelines for trustworthy AI | Shaping Europe’s digital future [European Commission of the EU]. Shaping Europe’s Digital Future. https://digital-strategy.ec.europa.eu/en/library/ethics-guidelines-trustworthy-ai (2019). Accessed 27 June 2022
European Commission: White Paper: On Artificial Intelligence—A European approach to excellence and trust (Communication from the Commission COM-65 final; White Paper, pp. 1–27). European Union. https://commission.europa.eu/system/files/2020-02/commission-white-paper-artificial-intelligence-feb2020_en.pdf (2020). Accessed 26 Jan 2023
Fersht, A.R.: AlphaFold–a personal perspective on the impact of machine learning. J. Mol. Biol. 433(20), 167088 (2021). https://doi.org/10.1016/j.jmb.2021.167088
Floridi, L.: AI and its new winter: from myths to realities. Philos. Technol. 33(1), 1–3 (2020). https://doi.org/10.1007/s13347-020-00396-6
Floridi, L.: The European legislation on AI: a brief analysis of its philosophical approach. Philos. Technol. 34(2), 215–222 (2021). https://doi.org/10.1007/s13347-021-00460-9
Floridi, L.: AI as Agency Without Intelligence: On ChatGPT, Large Language Models, and Other Generative Models (SSRN Scholarly Paper No. 4358789). https://doi.org/10.2139/ssrn.4358789 (2023)
Fulterer, R.: So spielt das Netz mit ChatGPT - und zeigt seine Grenzen auf. Neue Zürcher Zeitung. https://www.nzz.ch/technologie/so-spielt-das-netz-mit-chatgpt-ld.1716118 (2022). Accessed 23 Feb 2023
Furman, J., Seamans, R.: AI and the economy. Innov. Policy Econ. 19, 161–191 (2019). https://doi.org/10.1086/699936
Gabriel, I.: Artificial intelligence, values, and alignment. Mind. Mach. 30(3), 411–437 (2020). https://doi.org/10.1007/s11023-020-09539-2
Gao, L., Biderman, S., Black, S., Golding, L., Hoppe, T., Foster, C., Phang, J., He, H., Thite, A., Nabeshima, N., Presser, S., Leahy, C.: The pile: an 800GB dataset of diverse text for language modeling (arXiv:2101.00027). arXiv (2020). https://doi.org/10.48550/arXiv.2101.00027
Gille, F., Jobin, A., Ienca, M.: What we talk about when we talk about trust: theory of trust for AI in healthcare. Intell.-Based Med. 1, 100001 (2020). https://doi.org/10.1016/j.ibmed.2020.100001
Government of the Republic of Korea: Mid- to Long-Term Master Plan in Preparation for the Intelligent Information Society: Managing the Fourth Industrial Revolution (pp. 1–70) [Strategic Initiative]. Interdepartmental Exercise. https://k-erc.eu/wp-content/uploads/2017/12/Master-Plan-for-the-intelligent-information-society.pdf (2016). Accessed 26 Jan 2023
GPT-3: A robot wrote this entire article. Are you scared yet, human? The Guardian. https://www.theguardian.com/commentisfree/2020/sep/08/robot-wrote-this-article-gpt-3 (2020). Accessed 26 May 2022
Griffiths, T.L., Callaway, F., Chang, M.B., Grant, E., Krueger, P.M., Lieder, F.: Doing more with less: meta-reasoning and meta-learning in humans and machines. Curr. Opin. Behav. Sci. 29, 24–30 (2019). https://doi.org/10.1016/j.cobeha.2019.01.005
Hacker, P., Engel, A., Mauer, M.: Regulating ChatGPT and other large generative AI models (arXiv:2302.02337). arXiv (2023). https://doi.org/10.48550/arXiv.2302.02337
Häggström, O.: AI, orthogonality and the Muller-Cannon instrumental vs general intelligence distinction (arXiv:2109.07911). arXiv (2021). https://doi.org/10.48550/arXiv.2109.07911
Hahn, S.: KI-Startup Aleph Alpha macht USA und China mit eigenem KI-Basismodell Konkurrenz. Heise Online. https://www.heise.de/news/KI-Start-up-Aleph-Alpha-macht-USA-und-China-mit-eigenem-KI-Datenmodell-Konkurenz-6678538.html (2022). Accessed 26 May 2022
Haigh, T., Ceruzzi, P.E.: A New History of Modern Computing. MIT Press, Cambridge (2021)
Hernández-Orallo, J.: AI Safety Landscape from Short-Term Specific System Engineering to Long-Term Artificial General Intelligence. In: 2020 50th Annual IEEE/IFIP International Conference on Dependable Systems and Networks Workshops (DSN-W), pp. 72–73. https://doi.org/10.1109/DSN-W50199.2020.00023 (2020)
Hickman, E., Petrin, M.: Trustworthy AI and corporate governance: the EU’s ethics guidelines for trustworthy artificial intelligence from a company law perspective. Eur. Bus. Organiz. Law Rev. 22(4), 593–625 (2021). https://doi.org/10.1007/s40804-021-00224-0
Hinton, G., Sejnowski, T.J. (eds.): Unsupervised Learning: Foundations of Neural Computation, 1st edn. MIT Press, Boston (1999)
Hoffmann, J., Borgeaud, S., Mensch, A., Buchatskaya, E., Cai, T., Rutherford, E., Casas de L, D., Hendricks, L.A., Welbl, J., Clark, A., Hennigan, T., Noland, E., Millican, K., Driessche van den, G., Damoc, B., Guy, A., Osindero, S., Simonyan, K., Elsen, E., et al.: Training compute-optimal large language models (arXiv:2203.15556). arXiv (2022). https://doi.org/10.48550/arXiv.2203.15556
Hosseinia, M., Mukherjee, A.: Experiments with neural networks for small and large scale authorship verification (arXiv:1803.06456). arXiv (2018). https://doi.org/10.48550/arXiv.1803.06456
Huiling, E., Goh, B.: AI, robotics and mobility as a service: the case of Singapore. Field Actions Sci. Rep. Article Special Issue 17, 26–29 (2017)
Inria: TAILOR: A European Network for Trustworthy AI. https://www.inria.fr/en/tailor-european-network-trustworthy-ai (2022). Accessed 27 June 2022
Jannidis, F., Kohle, H., Rehbein, M.: Digital Humanities: Eine Einführung. J.B. Metzler (2017)
Jobin, A., Ienca, M., Vayena, E.: The global landscape of AI ethics guidelines. Nat. Mach. Intell. 1(9), 389–399 (2019). https://doi.org/10.1038/s42256-019-0088-2
Karnouskos, S.: Artificial intelligence in digital media: the era of deepfakes. IEEE Trans. Technol. Soc. 1(3), 138–147 (2020). https://doi.org/10.1109/TTS.2020.3001312
Kelleher, J.D.: Deep Learning (Illustrated Edition). The MIT Press (2019). Accessed 25 May 2022
Komeili, M., Shuster, K., Weston, J.: Internet-augmented dialogue generation (arXiv:2107.07566). arXiv (2021). https://doi.org/10.48550/arXiv.2107.07566
Kuchinke, B.A., Vidal, M.: Exclusionary strategies and the rise of winner-takes-it-all markets on the Internet. Telecommun. Policy 40(6), 582–592 (2016). https://doi.org/10.1016/j.telpol.2016.02.009
Kurz, S.: Digital Humanities: Grundlagen und Technologien für die Praxis. Springer-Verlag, Köln (2014)
Leahy, C.: Announcing GPT-NeoX-20B [H Report]. EleutherAI Blog. https://blog.eleuther.ai/announcing-20b/ (2022). Accessed 26 May 2022
Li, B., Qi, P., Liu, B., Di, S., Liu, J., Pei, J., Yi, J., Zhou, B.: Trustworthy AI: from principles to practices. arXiv (2021). https://doi.org/10.48550/arXiv.2110.01167
Lieber, O., Sharir, O., Lenz, B., Shoham, Y.: Jurassic-1: Technical Details and Evaluation (White Paper). AI 21 Labs, Tel Aviv (2021)
Lima, G., Cha, M.: Responsible AI and its stakeholders (arXiv:2004.11434). arXiv (2020). https://doi.org/10.48550/arXiv.2004.11434
Lin, J., Men, R., Yang, A., Zhou, C., Ding, M., Zhang, Y., Wang, P., Wang, A., Jiang, L., Jia, X., Zhang, J., Zhang, J., Zou, X., Li, Z., Deng, X., Liu, J., Xue, J., Zhou, H., Ma, J., et al.: M6: a Chinese multimodal pretrainer (arXiv:2103.00823). arXiv (2021). https://doi.org/10.48550/arXiv.2103.00823
Madiega, T.: EU guidelines on ethics in artificial intelligence: Context and implementation (EPRS PE 640.163; EU Human-Centric Approach to Artificial Intelligence, pp. 1–13). European Parliamentary Research Service. https://www.europarl.europa.eu/RegData/etudes/BRIE/2019/640163/EPRS_BRI(2019)640163_EN.pdf (2019). Accessed 26 Jan 2023
Massimov, K.: Künstliche Intelligenz: Masterplan für die Zukunft. Langen Mueller Herbig (2020)
Meskys, E., Kalpokiene, J., Jurcys, P., Liaudanskas, A.: Regulating Deep Fakes: Legal and Ethical Considerations (SSRN Scholarly Paper No. 3497144). https://papers.ssrn.com/abstract=3497144 (2019). Accessed 23 Jan 2023
Metzinger, T.: EU Guidelines: Ethics washing made in Europe [Section Politics]. Tagesspiegel Online. https://www.tagesspiegel.de/politik/ethics-washing-made-in-europe-5937028.html (2019). Accessed 26 Jan 2023
Michelucci, U.: Applied deep learning with TensorFlow 2: Learn to implement advanced deep learning techniques with Python. A-Press. https://doi.org/10.1007/978-1-4842-8020-1 (2022)
Microsoft: Reinventing Search with a New AI-Powered Microsoft Bing and Edge, Your Copilot for the Web. The Official Microsoft Blog. https://blogs.microsoft.com/blog/2023/02/07/reinventing-search-with-a-new-ai-powered-microsoft-bing-and-edge-your-copilot-for-the-web/ (2023). Accessed 23 Feb 2022
Ministry of Economic Affairs and Employment: Finland’s Age of Artificial Intelligence: Turning Finland into a leading country in the application of artificial intelligence (Turning Finland into a Leading Country in the Application of Artificial Intelligence, pp. 1–76) [Objective and recommendations for measures]. Finnish Government. https://julkaisut.valtioneuvosto.fi/bitstream/handle/10024/160391/TEMrap_47_2017_verkkojulkaisu.pdf (2017). Accessed 26 Jan 2023
Müller, M., Laurent, F.: Cedille: a large autoregressive French language model (arXiv:2202.03371). arXiv (2022). https://doi.org/10.48550/arXiv.2202.03371
Murakami, A., Thompson, P., Hunston, S., Vajn, D.: ‘What is this corpus about?’: Using topic modelling to explore a specialised corpus. Corpora 12(2), 243–277 (2017). https://doi.org/10.3366/cor.2017.0118
Muthukrishnan, N., Maleki, F., Ovens, K., Reinhold, C., Forghani, B., Forghani, R.: Brief history of artificial intelligence. Neuroimaging Clin. 30(4), 393–399 (2020). https://doi.org/10.1016/j.nic.2020.07.004
Narang, S., Chowdhery, A.: Pathways Language Model (PaLM): Scaling to 540 Billion Parameters for Breakthrough Performance [Research Report]. Google AI Blog. http://ai.googleblog.com/2022/04/pathways-language-model-palm-scaling-to.html (2022). Accessed 26 May 2022
Natale, S.: Deceitful Media: Artificial Intelligence and Social Life after the Turing Test. Oxford University Press, Oxford (2021)
National Program for Artificial Intelligence: UAE National Strategy for Artificial Intelligence (Strategic Initiative, pp. 1–25) [Outlook 2031]. United Arab Emirates. https://ai.gov.ae/wp-content/uploads/2021/07/UAE-National-Strategy-for-Artificial-Intelligence-2031.pdf (2018). Accessed 26 Jan 2023
National Science and Technology Council: The National Artificial Intelligence Research and Development Strategic Plan (pp. 1–48) [Networking and Information Technology Research and Development Subcommittee]. US Government. https://www.nitrd.gov/pubs/national_ai_rd_strategic_plan.pdf (2016). Accessed 26 Jan 2023
Nayak, P.: Understanding Searches Better than Ever Before [Research Report]. Google Blog. https://blog.google/products/search/search-language-understanding-bert/ (2019). Accessed 25 May 2022
Nijkamp, E., Pang, B., Hayashi, H., Tu, L., Wang, H., Zhou, Y., Savarese, S., Xiong, C.: A conversational paradigm for program synthesis (arXiv:2203.13474). arXiv (2022). https://doi.org/10.48550/arXiv.2203.13474
OECD Legal Instruments: OECD AI Principles (Recommendation of the Council on Artificial Intelligence OECD/LEGAL/0449; Committee on Digital Economy Policy, pp. 1–11). OECD. https://legalinstruments.oecd.org/en/instruments/OECD-LEGAL-0449 (2019). Accessed 26 Jan 2023
OpenAI: OpenAI Charter. Official Company Website. https://openai.com/charter/ (2018). Accessed 15 Apr 2022
OpenAI: OpenAI API: text-davinci-002 [Documentation]. GPT-3 Models. https://beta.openai.com (2022a). Accessed 26 May 2022
OpenAI: ChatGPT: Optimizing Language Models for Dialogue [Official Website]. Public Announcement. https://openai.com/blog/chatgpt/ (2022b). Accessed 23 Feb 2023
O’Regan, G.: A Brief History of Computing. Springer Nature, Mallow (2021)
Pichai, S.: An important next step on our AI journey [The Keyword]. Google News. https://blog.google/technology/ai/bard-google-ai-search-updates/ (2023). Accessed 23 Feb 2023
Preece, A., Harborne, D., Braines, D., Tomsett, R., Chakraborty, S.: Stakeholders in explainable AI (arXiv:1810.00184). arXiv (2018). https://doi.org/10.48550/arXiv.1810.00184
Qin, C., Zhang, A., Zhang, Z., Chen, J., Yasunaga, M., Yang, D.: Is ChatGPT a general-purpose natural language processing task solver? (arXiv:2302.06476). arXiv (2023). https://doi.org/10.48550/arXiv.2302.06476
Radford, A., Narasimhan, K., Salimans, T., Sutskever, I.: Improving Language Understanding by Generative Pre-Training [Research Report]. OpenAI. https://s10251.pcdn.co/pdf/2018-radford-gpt-1.pdf (2018). Accessed 26 May 2022
Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., Sutskever, I.: Language Models are Unsupervised Multitask Learners (Research Report) [OpenAI]. https://s10251.pcdn.co/pdf/2019-radford-gpt-2.pdf (2019). Accessed 26 May 2022
Rae, J.W., Borgeaud, S., Cai, T., Millican, K., Hoffmann, J., Song, F., Aslanides, J., Henderson, S., Ring, R., Young, S., Rutherford, E., Hennigan, T., Menick, J., Cassirer, A., Powell, R., van den Driessche, G., Hendricks, L.A., Rauh, M., Huang, P.-S., et al.: Scaling language models: methods, analysis and insights from training gopher (arXiv:2112.11446). arXiv (2022). https://doi.org/10.48550/arXiv.2112.11446
Reed, S., Zolna, K., Parisotto, E., Colmenarejo, S.G., Novikov, A., Barth-Maron, G., Gimenez, M., Sulsky, Y., Kay, J., Springenberg, J.T., Eccles, T., Bruce, J., Razavi, A., Edwards, A., Heess, N., Chen, Y., Hadsell, R., Vinyals, O., Bordbar, M., de Freitas, N.: A generalist agent (arXiv:2205.06175). arXiv (2022). https://doi.org/10.48550/arXiv.2205.06175
Renda, A.: Artificial Intelligence [Ethics, Governance and Policy Challenges]. CEPS. https://www.ceps.eu/ceps-publications/artificial-intelligence-ethics-governance-and-policy-challenges/ (2019). Accessed 26 Jan 2023
Roller, S., Weston, J., Dinan, E.: A state-of-the-art open source chatbot [Research Report]. Meta AI. https://ai.facebook.com/blog/state-of-the-art-open-source-chatbot/ (2020). Accessed 26 May 2022
Ruff, K.M., Pappu, R.V.: AlphaFold and implications for intrinsically disordered proteins. J. Mol. Biol. 433(20), 167208 (2021). https://doi.org/10.1016/j.jmb.2021.167208
Rumelhart, D.E., Hinton, G.E., Williams, R.J.: Learning representations by back-propagating errors. Nature (1986). https://doi.org/10.1038/323533a0
Saharia, C., Chan, W., Saxena, S., Li, L., Whang, J., Denton, E., Ghasemipour, S.K.S., Ayan, B.K., Mahdavi, S.S., Lopes, R.G., Salimans, T., Ho, J., Fleet, D.J., Norouzi, M.: Photorealistic text-to-image diffusion models with deep language understanding (arXiv:2205.11487). arXiv (2022). https://doi.org/10.48550/arXiv.2205.11487
Schlimbach, R., Asghari, R.: Das Digital Canvas: Ein Instrument zur Konzeption digitaler Geschäftsmodelle. HMD Praxis der Wirtschaftsinformatik 57(4), 866–878 (2020). https://doi.org/10.1365/s40702-020-00624-9
Shliazhko, O., Fenogenova, A., Tikhonova, M., Mikhailov, V., Kozlova, A., Shavrina, T.: mGPT: few-shot learners go multilingual (arXiv:2204.07580). arXiv (2022). https://doi.org/10.48550/arXiv.2204.07580
Smith, S., Patwary, M., Norick, B., LeGresley, P., Rajbhandari, S., Casper, J., Liu, Z., Prabhumoye, S., Zerveas, G., Korthikanti, V., Zhang, E., Child, R., Aminabadi, R.Y., Bernauer, J., Song, X., Shoeybi, M., He, Y., Houston, M., Tiwary, S., Catanzaro, B.: Using DeepSpeed and megatron to train megatron-turing NLG 530B, a large-scale generative language model (arXiv:2201.11990). arXiv (2022). https://doi.org/10.48550/arXiv.2201.11990
Smuha, N.A.: The EU approach to ethics guidelines for trustworthy artificial intelligence. Comput. Law Rev. Int. 20(4), 97–106 (2019)
Strategic Council for AI Technology: Artificial Intelligence Technology Strategy [Report of Strategic Council for AI Technology)]. Japanese Government (2017)
Tafjord, O., Clark, P.: General-purpose question-answering with Macaw (arXiv:2109.02593). arXiv (2021). https://doi.org/10.48550/arXiv.2109.02593
TAILOR Initiative: TAILOR Objectives [Official Website]. The TAILOR Network. https://tailor-network.eu/about/tailor-objectives/ (2022). Accessed 27 June 2022
Tay, Y., Dehghani, M., Tran, V.Q., Garcia, X., Bahri, D., Schuster, T., Zheng, H.S., Houlsby, N., Metzler, D.: Unifying language learning paradigms (arXiv:2205.05131). arXiv (2022). https://doi.org/10.48550/arXiv.2205.05131
Technology Innovation Institute: Technology Innovation Institute Announces Launch of NOOR, the World’s Largest Arabic NLP Model [Research Report]. TII News. https://www.tii.ae/news/technology-innovation-institute-announces-launch-noor-worlds-largest-arabic-nlp-model (2022). Accessed 26 May 2022
Thompson, A. D.: First look—Imagen by Google AI/Research/Brain—Launched May/2022—(2B + 1B + 4.6B T5-XXL) [Research Update]. Life Architect. https://www.youtube.com/watch?v=NYGdO5E_5oY (2022a). Accessed 26 May 2022
Thompson, A. D.: Timeline of AI and Language Models [Science Report]. Life Architect. https://lifearchitect.ai/timeline/ (2022b). Accessed 26 May 2022
Thoppilan, R., De Freitas, D., Hall, J., Shazeer, N., Kulshreshtha, A., Cheng, H.-T., Jin, A., Bos, T., Baker, L., Du, Y., Li, Y., Lee, H., Zheng, H.S., Ghafouri, A., Menegali, M., Huang, Y., Krikun, M., Lepikhin, D., Qin, J., et al.: LaMDA: language models for dialog applications (arXiv:2201.08239). arXiv (2022). https://doi.org/10.48550/arXiv.2201.08239
Tong, Z., Zhang, H.: A Text Mining Research Based on LDA Topic Modelling. In: Computer Science and Information Technology (CS and IT), pp. 201–210. https://doi.org/10.5121/csit.2016.60616 (2016)
Turing, A.M.: 1. Computing machinery and intelligence. Mind LIX(236), 433–460 (1950). https://doi.org/10.1093/mind/LIX.236.433
Turkbey, B., Haider, M.A.: Artificial Intelligence for automated cancer detection on prostate MRI: opportunities and ongoing challenges, from the AJR special series on AI applications. Am. J. Roentgenol. (2021). https://doi.org/10.2214/AJR.21.26917
Underwood, T.: Dear Humanists: Fear Not the Digital Revolution [The Review]. The Chronicle of Higher Education. https://www.chronicle.com/article/dear-humanists-fear-not-the-digital-revolution/ (2019). Accessed 30 June 2022
United Nations: A United Nations system-wide strategic approach and road map for supporting capacity development on artificial intelligence (First Regular Session 2019 CEB/2019/1/Add.3; Summary of Deliberations, pp. 1–12). UN. https://unsceb.org/sites/default/files/2020-09/CEB_2019_1_Add-3-EN_0.pdf (2019). Accessed 26 Jan 2022
United Nations: Artificial Intelligence and Robotics [Official Communication]. UN AI and Robotics. https://unicri.it/index.php/topics/ai_robotics (2020). Accessed 26 Jan 2022
Uszkoreit, J.: Transformer: A Novel Neural Network Architecture for Language Understanding [Google Research]. Google AI Blog. http://ai.googleblog.com/2017/08/transformer-novel-neural-network.html (2017). Accessed 25 May 2022
van der Aalst, W., Hinz, O., Weinhardt, C.: Big digital platforms. Bus. Inf. Syst. Eng. 61(6), 645–648 (2019). https://doi.org/10.1007/s12599-019-00618-y
Vuong, Q.-H., Ho, M.-T., Vuong, T.-T., La, V.-P., Ho, M.-T., Nghiem, K.-C.P., Tran, B.X., Giang, H.-H., Giang, T.-V., Latkin, C., Nguyen, H.-K.T., Ho, C.S.H., Ho, R.C.M.: Artificial intelligence vs. natural stupidity: evaluating AI readiness for the Vietnamese medical information system. J. Clin. Med. (2019). https://doi.org/10.3390/jcm8020168
Walter, Y.: Building human systems of trust in an accelerating digital and AI-driven world. Front. Hum. Dyn. (Sect. Digit. Impacts) 4(926281), 1–5 (2022). https://doi.org/10.3389/fhumd.2022.926281
Walter, Y.: Künstliche Intelligenz und datenbasierte Geschäftsmodelle [Kalaidos FH Blog]. Digital Ventures. https://www.kalaidos-fh.ch/de-CH/Blog/Posts/2022/04/Digitalisierung-1069-Kuenstliche-Intelligenz-und-datenbasierte-Geschaeftsmodelle (2022b). Accessed 29 June 2022
Walter, Y.: Der digitale Mensch in der Gesellschaft und Wirtschaft [Kalaidos FH Blog]. Digital Ventures. https://www.kalaidos-fh.ch/de-CH/Blog/Posts/2022/05/Digitalisierung-1070-Digitaler-Mensch-in--Gesellschaft-und-Wirtschaft (2022c). Accessed 29 June 2022
Wang, B.: GTP-J-6B [Python; GitHub]. https://github.com/kingoflolz/mesh-transformer-jax (Original work published 2021) (2021). Accessed 29 June 2022
Webster, G., Creemers, R., Kania, E., Triolo, P.: Full Translation: China’s ‘New Generation Artificial Intelligence Development Plan’ (Academic Translations) [DigiChina]. Stanford University. https://digichina.stanford.edu/work/full-translation-chinas-new-generation-artificial-intelligence-development-plan-2017/ (2017). Accessed 26 Jan 2023
Weerasinghe, J., Greenstadt, R.: Feature Vector Difference based Neural Network and Logistic Regression Models for Authorship Verification. CEUR Workshop Proceedings, 2695. https://par.nsf.gov/biblio/10249189-feature-vector-difference-based-neural-network-logistic-regression-models-authorship-verification (2020). Accessed 30 June 2022
Wirtz, B. W.: Digital Business Models: Concepts, Models, and the Alphabet Case Study. Springer, London (2019)
Wooldridge, M.: A Brief History of Artificial Intelligence: What It Is, Where We Are, and Where We Are Going. Flatiron Books, New York (2021)
Wu, S., Zhao, X., Yu, T., Zhang, R., Shen, C., Liu, H., Li, F., Zhu, H., Luo, J., Xu, L., Zhang, X.: Yuan 1.0: large-scale pre-trained language model in zero-shot and few-shot learning (arXiv:2110.04725). arXiv (2021). https://doi.org/10.48550/arXiv.2110.04725
Yuan, S., Zhao, H., Du, Z., Ding, M., Liu, X., Cen, Y., Zou, X., Yang, Z., Tang, J.: WuDaoCorpora: a super large-scale Chinese corpora for pre-training language models. AI Open 2, 65–68 (2021). https://doi.org/10.1016/j.aiopen.2021.06.001
Zhang, H., Li, L.H., Meng, T., Chang, K.-W., den Broeck, G.V.: On the paradox of learning to reason from data. arXiv (2022). https://doi.org/10.48550/ARXIV.2205.11502
Zhang, S., Roller, S., Goyal, N., Artetxe, M., Chen, M., Chen, S., Dewan, C., Diab, M., Li, X., Lin, X.V., Mihaylov, T., Ott, M., Shleifer, S., Shuster, K., Simig, D., Koura, P.S., Sridhar, A., Wang, T., Zettlemoyer, L.: OPT: open pre-trained transformer language models (arXiv:2205.01068). arXiv (2022). https://doi.org/10.48550/arXiv.2205.01068
Zhuang, S., Hadfield-Menell, D.: Consequences of misaligned AI. Adv. Neural. Inf. Process. Syst. 33, 15763–15773 (2020)
Zhuo, T.Y., Huang, Y., Chen, C., Xing, Z.: Exploring AI ethics of ChatGPT: a diagnostic analysis (arXiv:2301.12867). arXiv (2023). https://doi.org/10.48550/arXiv.2301.12867
Zou, X., Yin, D., Zhong, Q., Ding, M., Yang, H., Yang, Z., Tang, J.: Controllable Generation from Pre-Trained Language Models via Inverse Prompting. In: Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pp. 2450–2460. https://doi.org/10.1145/3447548.3467418 (2021)
Funding
Open access funding provided by Kalaidos University of Applied Sciences.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The author has no conflicts of interest to declare.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Walter, Y. The rapid competitive economy of machine learning development: a discussion on the social risks and benefits. AI Ethics 4, 635–648 (2024). https://doi.org/10.1007/s43681-023-00276-7
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s43681-023-00276-7