
Writing at a Distance: Notes on Authorship and Artificial Intelligence
This essay provides an overview of notions of authorship in AI and natural language processing systems and discusses past and current debates about computers as literary authors. It suggests the concept of causal authorship to measure the types of distance between human and machine agents, acknowledging the anthropocentric bias of this idea. To balance this, the essay reflects on the notion of distributed authorship, which considers the network of actors involved in the creation of a text and has its own limitations. Both concepts are elements of a future theory of authorship in the age of machine learning.
With the advent of OpenAI’s ChatGPT, Anthropic’s Claude, Google’s Gemini, and other text-generating machine learning systems, the old boundaries between the “two cultures” of the humanities on the one hand and what is now called STEM on the other seem to be fraying by the day.1 As “large language models” become more sophisticated and capable of producing writing that is at least in certain cases indistinguishable from human-written text, it is less and less clear that the instance of writing can be so easily located. This is not only a challenge for literary studies, for which the discussion of the “author function” is one of its core competencies.2 Engineering—the “E” in STEM—in the form of artificial intelligence research, too, must ponder this question, since the goal of a strong AI presupposes that computers are autonomous agents, and as such can potentially claim authorship.3 And while literary studies have traditionally sought to deconstruct the singular and autonomous author-genius, current fantasies about AGI (artificial general intelligence) perform a return of Romantic ideas of geniuses in machinic form.4
Authorship, in other words, is now a benchmark for computer science, as evidenced by collaborative programs like Google’s Artists and Machine Intelligence.5 [End Page 315] However, what appears novel often has a long history, and computer-generated language is almost as old as computers themselves.6 Natural language processing (NLP) can be traced back to the 1940s,7 when Warren Weaver’s memorandum on the need for machine translation opened up a promising field of research in the young discipline of computer science,8 which has often been cited as the “birth” of automated language processing.9
Even then, the question of authorship of synthetic text played an unacknowledged role in the broader field of artificial intelligence. Alan Turing’s famous 1950 “imitation game,” later known as the “Turing test,” was a pivotal moment in this regard. Turing argued, in good behaviorist fashion, that in order to test whether a machine was intelligent, it had to be able—at least in terms of its conversational abilities—to fool a human into thinking that it, too, was a human.10 Since for technical reasons this conversation would have to be textually mediated, this means that “artificial intelligence is the art of making texts,” as Jay David Bolter put it already back in 1991.11 Authorship has thus been a central concern in AI research from the very beginning, since any text generated by a machine must convey a sense of intentionality if it is to be considered intelligent.
Until recently, however, the potential of AI-generated language remained largely untapped due to the fact that AI text production was primarily situated outside literary contexts. The users of Joseph Weizenbaum’s 1964 ELIZA program, for example—the earliest interactive chatbot, which simulated a Rogerian psychoanalyst—accepted the computer against their better judgment as an intentional communication partner; they famously began having serious and private conversations with it.12 But it would be incorrect to say that they also saw ELIZA as an author in the specifically literary meaning of the word. The production of explicitly literary texts—and the associated distinction of a specifically literary “mode of being of discourse” as well as the “classificatory function” of work coherence and intellectual ownership,13 for which the term authorship is normally reserved in its emphatic meaning—was simply not a priority for mainstream AI development.14 Instead, these ideas were explored mainly in the artistic-experimental fringes of NLP.15
This essay takes up the parallelism of intentionality and authorship on the one hand and NLP and AI on the other in order to give, in the first section, an overview of past and present debates about computer authorship, and to show some continuities that connect them to the more traditional discourse of literary studies, as well as the surprisingly old discussion about the artistic use of the computer. Because the line between “AI” and mere computer use is often difficult to draw, I will briefly summarize the historical debates about authorship of computer-generated literary texts before systematically addressing authorship specific to the narrower field of AI; many of the themes of the former recur in the latter.
It would, however, be wrong to assume that there is absolutely nothing new under [End Page 316] the sun. In particular, the shift to unsupervised statistical reasoning as the base technology of AI research is a genuinely novel phenomenon that requires novel concepts for describing authorship in and around these systems. The second part of this essay thus introduces the idea of causal authorship as a measure of the distance between human and machine agent that has no obvious parallel in the history of authorship.
Yet causal authorship is only a heuristic tool and still thinks of the human at the end of the chain of operations as the author. Thus, the third section attempts to counter this overly anthropocentric scheme with a notion of distributed authorship that is spread across a network of many actors; this model, however, has its own problems as it makes it more difficult to address political and ethical concerns. Both notions of authorship, then, should be seen as checking each other, and are open to being extended. For what I present here is after all only a first sketch. It expresses the need to rethink authorship in the age of large language models and while it leaves many questions open, it can serve as a basis for a more in-depth exploration of the topic.
Historical Outline of Computer-Generated Literature
Christopher Strachey, a colleague of Turing’s at Manchester University, is credited with creating the first literary experiment using a computer. His “Love Letters” were produced in 1952 on the Manchester Mark I and printed out via teletype. These letters were generated through a combinatorial process in which a set of fixed words was inserted into a variety of possible word patterns for a love letter,16 each of which was signed “M.U.C.” (Manchester University Computer).
Already here, the computer is presented as an author—but only in the sense of a playful authorship fiction, hardly very different from that of E. T. A. Hoffman’s tomcat Murr. Strachey never claimed that the computer actually assumed the role of the author, but rather saw such notions as an anthropomorphic projection. He stated that the question was not whether computers were capable of writing letters, but rather whether one could write a program for that purpose, “and this is really a very different sort of question because the writing of programs . . . is still an essentially human activity.”17 Authorship therefore rested with the programmer, not the computer. Like Turing, however, Strachey considered the possibility of successfully deceiving readers—and thus implicitly the assumption of computer authorship—not only possible but probable in the future.18
In the German context, the information aesthetics of the Stuttgart School, centered around philosopher and writer Max Bense, introduced a new perspective on authorship that complicated previously established ideas by pointing out their ontological assumptions. While Strachey, following Turing’s proposal, emphasized the appearance of intentionality as a precondition of authorship—and recommended clever programming as a way of bringing about this deception—information aesthetics articulated a break in the ontology of the text itself. [End Page 317]
In his essay “Über natürliche und künstliche Poesie” (“On Natural and Artificial Poetry”), Bense ascribes to the former “an ego relation and a world aspect” that continue to operate in a text as an “ontological continuation” of personal or social experiences of the world. By contrast, “artificial poetry” has, in principle, no intentional and subject-relative, but only a “material origin.” It consists of contextless textual elements that lack any human meaning, but whose aesthetic content can be independently measured on the basis of communication theory and the statistical improbability of symbol sequences.19
This view relativizes the question of authorship by prioritizing text-immanent criteria over the author’s identity. Understood this way, computer-generated texts lack intentionality by definition, rendering the Turing test a category mistake. The fact that Bense could speak of “simulated poetry”20 highlights this ontological distinction rather than eliminating it, aligning his text theory with the discourse on the “death of the author” of Roland Barthes, Michel Foucault, and Umberto Eco.21
The focus on the quantitative properties of text, such as word frequency and vocabulary distribution, is also evident in Theo Lutz’s “Stochastic Texts” of 1959, regarded as the first computer-generated literature in the German-speaking world. Lutz, a student of Bense’s, tasked a program to combine a list of words into short sentences by a weighted random process. The fact that the vocabulary used came from Kafka’s Schloß does not contradict the devaluation of the authorship question.22
Lutz, however—less interested in theory than Bense—seems to have related his own authorship primarily to the program script, while treating the output only as its secondary effect.23 As with Strachey, Lutz’s work is situated in the realm of NLP, not AI; he explicitly declares it to be the synthetic counterpart of an “analytic philology.”24 Something similar can be said of the linguist Gerhard Stickel’s “autopoems,” generated from 1965 onward, which function according to a comparable pattern. Stickel, too, considered the assumption of computer authorship to be an inadmissible anthropomorphization, “justified only by ignorance of its mode of operation.”25
What these two examples show is that it is not the search for autonomously acting, problem-solving or even consciousness-equivalent machines, but the analysis and synthesis of syntactic structures that appears as the starting point of computer-generated literature. This setting, however, deprivileges the authorship question quite pragmatically (in the case of Lutz and Stickel) or through a strong notion of text on a theoretical level (as in the case of Bense). Where the linguistic or computer science approach increasingly gives way to an artistic one, it seems that in many cases the poetological interest in the “death” of the human author has also prevented its resurrection as a computer author.26
This is evident in the generative texts of the 1960s and 1970s, which tended to be produced in the context of modernist aesthetics that experimented with aleatoric processes in the wake of John Cage and Marcel Duchamp, rather than projecting [End Page 318] authorship, let alone “genius,” onto machines. The relevant works of the period, at least, seem to indicate as much. The French group OuLiPo (L’Ouvroir de littérature potentielle; workshop of potential literature), operating with self-imposed constraints, was from the start interested in mathematized text creation; its member Jean Baudot published La machine à écrire (The writing machine) in 1967, a collection of free verse generated with his software PHRASE.27 Austrian poet Marc Adrian’s “Maschinentexte,” begun in the late 1950s and working with permutations and loops, stem in a direct line from the linguistic avant-gardism of the Vienna group around Gerhard Rühm and H. C. Artmann.28 In the case of Alison Knowles and James Tenney’s “A House of Dust” (1968), the collaboration between art (Knowles) and engineering (Tenney) is also reflected in the equality of the authorial duo.29 But when Dick Higgins, who like Knowles belonged to the Fluxus movement, argued for the use of “computers for the arts” two years later, he continued to do so under the proviso that “Computers are like most tools—deaf, blind and incredibly stupid.”30
Whenever a conscious staging of authorship occurs, it is hardly taken very far at this point: for the collection of Computer-Lyrik (1968)—comprising poems that emulate both classical authors as Goethe and contemporary poets like Paul Celan—the programmers Manfred Krause and Götz F. Schaudt merely signed as editors, thus conceding authorship to the machine on a paratextual level; in the preface, however, this gesture is retracted and the computer is referred to as “only a tool in our hand.”31 The poet Karl Krolow, who reviewed the volume the same year, nevertheless played along and, with tongue-in-cheek seriousness, ascribed authorship to “Zuse Z 23, one-year-old computer poet.”32
The discrepancy between the production and reception of computer literature is striking: for many critics, it serves as a foil for man-machine comparisons—either indicating an irreducible human faculty for “real” literature, as Italo Calvino locates this capacity in the rootedness of every story in myth,33 or suggesting a mere measure of craft, as Hans Magnus Enzensberger put it in 1974: “Whoever cannot write poetry better than the machine ought rather not to do it at all.”34 Despite this repetition of technocritical tropes,35 the discussion of authorship remained conspicuously secondary for those producing such literature.
Austrian authors Franz Josef Czernin and Ferdinand Schmatz developed the rule-based poetry generator POE (“Poetic Oriented Evaluations”) in the late 1980s, which analyzed the character distribution of a text, including its phonetic qualities, and allowed transformations and substitutions. Thus, Hölderlin’s poem “Andenken,” with the first line “Wo aber sind die Freunde? Bellarmin” could be transformed into the phonetically similar line “sog aber licht, die leuchten wellenartig” in Czernin/Schmatz’s poem “tag.”36 POE, too, was only “meant to be an assistant,” as Czernin recently, and with an eye to ChatGPT, stressed in an interview: “We were explicitly not interested in the machine producing works itself.”37 [End Page 319]
This attitude however also changed in the 1980s with the publication of the book The Policeman’s Beard is Half Constructed (1984), in which the staging of autonomous machine authorship is taken to an unprecedented extreme: advertised on the cover as “the first book ever written by a computer,” the program “Racter” paratextually figures as the author.38 The lavish book edition, complete with images that look like a cross between stills from the movie Tron and the xylographs of Max Ernst, no longer stands for a playful, transparent fiction of computer authorship but arguably for its serious assertion.
William Chamberlain and Thomas Etter, the developers of Racter (although Etter himself may have been fictional), have since come under some criticism for neither publishing the program code nor revealing the depth of their editorial intervention. As a result, The Policeman’s Beard is now considered a highly controversial key work of generative literature.39 Significantly, the back cover of the book distinguishes Racter from AI: “Fundamentally different from artificial intelligence programming, which attempts to replicate human thinking, Racter can write original work without promptings from a human operator.” Autonomous machine authorship is declared here to transcend the paradigm of AI, which should be seen not least in the light of an AI discourse that regained momentum in the 1980s.
Indeed, the 1980s were something of a crossroads. The development of text-based AI and generative writing had long run in parallel. Weizenbaum’s ELIZA had shown that a computer could pass the Turing in the sense that readers can project agency onto it. However, literary interest in text generators waned in the late 1980s and 1990s. Instead, an entirely different form of computer-centered writing emerged in the form of hypertext literature, which privileged modes of reception rather than production, and sent readers down “forking paths.”40 In contrast to early NLP, which was entirely removed from the high theory of the 1970s and 1980s, hypertext literature can be considered poststructuralism turned technology: since hypertext structures allow for multiple reading paths, they not only replace the authoritarian “tree” with the anarchic “rhizome,”41 they also strip the author of his or her power, who is now a mere scripteur of collections of material,42 and valorize the reader as the actual meaning-making agent. New figures were proposed, such as the “wreader”43 or the “browser”,44 who forge their own nonlinear paths through the linked sections of text.45
However, the insistence on the “networking aspect”46 and the “interactivity, intermediality, and staging”47 of reception, on which scholars of the time focused, tended to obscure the conditions of production; especially in the German context, Netzliteratur (net literature) was for a long time confusingly used as the generic term for all computer-based and thus also generative literature, even if it had nothing to do with the net. Today, however, the then-novel focus on collaboration is once again central to the discussion of authorship in AI: just as different authors work together in large Mitschreibeprojekte (co-writing projects),48 the relationship between humans [End Page 320] and machines could also be described as cooperative, as “cyborg authorship.”49 Current approaches, such as those of the German writer Mattis Kuhn, the Austrian digital poet Jörg Piringer, and myself,50 increasingly take up this aspect again, and I will come back to it in the third section.51
Causal Authorship
This—admittedly condenses—overview of the discourse on machine authorship shows how NLP and AI research, as well as artistic approaches, articulate quite different interests: while AI research, with its hunt for intelligence or rationality (if not consciousness), continues to at least implicitly presuppose authorship as an intentional category, NLP is more interested in the processes of text synthesis, which, depersonalized and as structural property, can be understood as a system of langue entirely without parole, to speak with Saussure. Artistic experiments can make use of both frameworks and are driven much more by aesthetic and poetological than by technical considerations, so that the discourse of authorship has received a completely different technical basis with the poststructuralist-influenced hypertext literature.
The neglect of the production aspect in the literary theoretical discourse of the 1980s and 1990s, however, overlooked the fact that a fundamental reorientation of AI research took place during this period, which, with some interruptions and detours, is still influential today. If the focus had been primarily on symbolic AI since the field’s inception at the Dartmouth Conference in 1956, the 1980s saw a shift to subsymbolic AI.52 The difference is fundamental.
Symbolic AI involves the explicit formulation of rule steps as well as the encoding of world knowledge in so-called expert systems.53 These systems, programmed according to a “sequential” paradigm in a step-wise fashion, are human-readable in their structure, and their ability to learn is limited to expanding their knowledge base. Subsymbolic AI, on the other hand, whose foundations were laid in the 1950s but not revisited until the 1980s,54 does not require explicit rules but is based on stochastic machine learning. It can be called the “connectionist” paradigm, since deep learning primarily involves artificial neural networks (ANNs) that model an input-output function via amplifying and inhibiting the signals of “neuron” structures, that, abstractly, emulate the structure of the brain.55
Nevertheless, it took until the 2010s for the connectionist paradigm to be used for literary and artistic purposes. Since then, its power has been demonstrated in comparison to classical text generators based on combinatorics or aleatorics. Moreover, in unsupervised learning—in which the features of language are derived by the machine learning system without any human assistance—the idea of “intelligent” machines is realized much more convincingly, so that they can apply and accumulate knowledge with greater ease. Since current NLP applications such as machine translation are also implemented as ANNs and, moreover, are increasingly referred [End Page 321] to as “AI” themselves, the former distance between the two fields has shrunk. “Large language models” such as Google’s Gemini are now being presented as “foundation models” on which more specialized AI systems are set up.56
The first explicit literary experiments with ANNs may have taken place in the early 2000s.57 However, it was the popularization of an AI architecture called LSTM-RNN (long short-term memory recurrent neural network) around 2015—although they had been known since the 1990s—that spawned the broad field of literary experiments that continues to flourish today.58 As the quality of the output (in terms of approximating natural language) has improved even more since then—the currently most successful architecture is the Transformer, sometimes with an added layer of human feedback optimization—the question of the authorship of such systems has come back into focus.59 Significantly, the technical substructure has a direct impact on the possible loci of authorship. It makes a difference whether writing is sequential or connectionist, as the chain of operations and technical apparatuses between the authoring human and the final text change when moving from one to the other.
To describe these different configurations, I propose the concept of causal authorship (fig. 1). Before a specifically literary author function is concerned, we must first be clear about who does what by which means at which point in the process of text production. Looking in this way at the production side of generative writing, causal authorship constructs degrees of distance between the human author and resultant text. This model is willfully reductive—it assumes a single author, only looks at the process of text generation (rather than, say, brainstorming prior to writing), ignores the historical changes in the concept of authorship, and it focuses solely on the reception side—but in its constructedness shows how the writing configuration fundamentally changes with new technologies.
Primary authorship would be the still valid standard of conventional, “immediate” writing, in which an author more or less directly puts text on paper or onto a data carrier. The entities that stand between the writer and the text may be technical in nature—from quill to typewriter to word processor—and may have some effect on the output, as media studies has repeatedly pointed out, but their determinative influence does not diminish the connection between author and text that is perceived as causally unbroken.60 The act of writing is localized entirely or predominantly on the side of the person operating the apparatus of writing, and all other actors involved are merely passively mediated, or have a maybe measurable but in the end non-decisive impact on the author. That this model is by no means valid for all forms of writing can be seen by turning to the additional degrees of distance that rupture this unbroken connection and insert a gap in its relative immediacy.
Secondary authorship would thus be a first-order distance that duplicates the act of writing. Strictly speaking, it appears in all rule-based poetry, be it the combinatorial poetry experiments of the Baroque or those of the historical avant-gardes.61 It is also [End Page 322] the model of authorship at work in the sequential paradigm: the author’s contribution in these cases is to formulate a sequence of rules, the execution of which produces the work. The act that establishes this rule sequence is, on a causal level, indistinguishable from primary authorship, but the resulting text has a vicarious function, being merely the first module in a chain of operations that outputs the final text. Authorship here means writing a text that writes a text.
Four degrees of causal authorship measuring the “distance” between human author and final text (figure by Author).
This model includes all of the types of classic digital literature mentioned above, such as Strachey’s “Love Letters” (though hypertext is subject to certain limitations), as well as the transformation rules of symbolic AI, as in ELIZA. Unlike historical rule-based poetry, however, the text of the code is itself operative—it has its own causality and performativity as it effectively brings about the output.62 It is this performative relationship that raises the question of a possible computer authorship in the first place, which is hardly an issue in “analog” rule-based writing.
However, it is not always necessary that the operative text be human made; it may itself be the product of an earlier production step. The machine learning process of the connectionist paradigm inserts a second hiatus into the chain of operations and brings forth tertiary authorship. Computer scientist François Chollet has described the essential difference between the sequential and the connectionist paradigms as follows: if the former can be described as “rules + data = answers,” the latter follows the scheme “data + answers = rules.”63 Thus, rules are no longer written in a program script whose application (to data) produces an output, but rather an ANN is trained on a large set of exemplary outputs that makes the “rules” that eventually lead to the final text. The exact model architecture is not important for determining [End Page 323] the degree of distance; the principle is the same for LSTM-RNNs or the Transformer: it is up to the author to program the learning algorithm (which is, however, usually done by third parties), to define the training dataset (from which the ANN builds the model on its own), and to determine the parameters (by which the model finally produces the output). The text that writes the output in the secondary model is thus itself computer-made.
Nevertheless, even in the tertiary model, authorship is still located, as it were, in the engine room of the digital: the choice of the training algorithm and the “hyper-parameters”—settings such as the learning rate of the model or the randomness (“temperature”) of the output—as well as the curation of the training dataset are still subject to the user’s control and choice. With the advent of large language models such as ChatGPT, Claude, or Gemini,64 this may not always be the case. One can now plausibly speak of quaternary authorship. Since large language models are predominantly proprietary software—which is still in most cases too large to be trained from scratch by individual users—consumers are more often than not limited to the “factory settings” and cannot choose their own dataset.65 What is more, the only way to interact with these models is through their graphical interface or their API via natural language prompts. Thus, instead of writing the code that writes a poem, or training a language model on a large number of poems and having it output new ones, in the quaternary model, it is solely the input that counts and that is still under human control: “Write a poem in the style of Wallace Stevens.”66 “Promptology”—the efficient, even virtuosic formulation of such input prompts—is the main mode of operation of quaternary authorship.67 But because it can only operate within a given system, it also means: Il n’y a rien hors du modèle—there is no outside-model. Authorship is encapsulated in the iron cage of a commercial language technology whose precise workings are unfathomable to the outside user.68
These degrees of causal authorship are ideal types, intended to bring some structure to what has so far been a rather unstructured debate about AI authorship. They are, therefore, not exhaustive, nor are they meant to indicate any kind of teleology or progression from low to high. In particular, the latter two are in practice simply a choice between using the “factory settings” or not—regardless of whether fine-tuning via one’s own dataset is possible, as with the current GPT-3 model, or not, as with Gemini. K Allado-McDowell’s Pharmako-AI, which the author wrote using the standard GPT-3 setting, would thus be quaternary writing; training an AI model on a specific corpus, as I did for my novel (Berlin, Miami), would be tertiary writing.69 The ordinal numbering does not declare one type of writing to be more advanced than the other, only that the distance from the author to the final text, and thus the operative choices open to the human author, differ between them.
There are certainly cases that complicate this ideal-typical taxonomy. One would be the kind of electronic literature in which the programming script is itself a readable [End Page 324] poem but also produces one when it is executed, as in the case of “Black Perl” code poetry.70 Here, we have a simultaneity, or entanglement, of primary and secondary authorship. Another problematic case is the status of a program’s appropriated output. An example here is Peter Dittmer’s artwork Die Amme (The wet nurse), which he first realized in 1992 and which has seen several versions since.71 The complex art installation features, among other things, a robotic arm that would sometimes tip over a glass of milk, but also a number of terminals that allow visitors to converse with the Amme. In a manner similar to Weizenbaum’s ELIZA, it would react to the input via a large set of symbolic transformation rules. In 2010, the poet Ulf Stolterfoth published a series of conversations he had with the Amme with poetry publisher Urs Engeler.72 The static text, rather than the dynamic conversation, now invokes the dialog as a literary genre instead of the lived activity that would give rise to questions of intentionality and intelligence. Yet while Dittmer, as programmer, can count as having a relation of secondary authorship to the output (excerpts of which he also published), Stolterfoth’s approriation—the product of a mere interface interaction with the machine—is either a sequential form of quaternary authorship or something entirely different.
Distributed Authorship
Causal authorship, which focuses on the human component in human-machine interactions and measures its distance from the final text, is a necessary first step toward more complex descriptive models of authorship in the age of large language models. Identifying all the technical processes involved is a prerequisite for moving on to more advanced conceptualizations. Instead of addressing the limitations of this model that arise from its idealizations, I now want to consider two that are more concerned with its underlying ontology of authorship itself: first, it seems already decided here that computers cannot be authors strictly speaking but are only cogs in a larger machinery of writing that is under the control of a human, however distant he or she may be. And second, it is still a single author who is ultimately responsible for the resulting text, and not a whole network of actors, be they human or non-human.
To begin with, it is not so certain that the human is always in the driver’s seat when writing a text. Since primary authorship has so far served as a foil for “immediate” writing, the increasing distance seems to be accompanied by an increasing loss of control: while the production rules of secondary authorship in the sequential paradigm could still be formulated precisely and deterministically, the machine learning approach of the connectionist paradigm that governs tertiary and quaternary authorship operates via a stochastic process that results in a probability distribution over text sequences. If the former can easily produce any number of identical outputs, the latter is designed to reproduce statistically similar, but not identical outputs; instead of slot-and-fill templates, it outputs “Gestalten.”73 [End Page 325]
In addition, neural language models are largely opaque. While the code that characterizes secondary authorship is written by humans and can therefore also be read by them, the “weight model” of the ANN—which encodes the connections strengths between its simulated neurons—is, as the result of a statistical process, in principle not translatable back into human-readable rule-steps of code.74 In Diltheyian terms, the ANN is not straightforwardly an artifact of human cultural production to be understood, but rather akin to a natural object to be explained—a star cluster rather than a manuscript variant.75
Both the loss of control and the increasing opacity of language models raise the possibility of autonomous AI authorship—not the human but the machine itself would be the author here. The essential axis, then, is that of the agency of the system, which ranges from mere tool to autonomous agent: is AI itself an author in the end or just a better typewriter?76
As shown in the first section, classical computer science, as well as the bulk of digital literature, resists attributing authorship to the computer. Weizenbaum, for instance, still held in 2001: “The programmer has chosen the rules according to which the computer writes. Whether now the programmer deserves praise because he has chosen just these editing rules, I don’t know. . . . In any case, he should not say that the computer writes poetry.”77 But there have always been counter-arguments. Already Turing had considered Weizenbaum’s point under the title of “Lady Lovelace’s objection.”78
He was alluding to Ada Lovelace, who developed the “programming language” for Charles Babbage’s mechanical calculating machine and who can thus be considered the first coder. In a commentary on the design of Babbage’s analytical engine, she speculated that the machine could not, in principle, produce anything new because all of its products could be traced back to the instructions of a human programmer.79 Turing, on the other hand, held that even rule-based programs could produce surprise because human anticipation is limited in the face of complex code-output relationships.80 Moreover, the possibility of learning machines would render Lovelace’s objection moot: if a machine were capable of dynamically adopting new behavior on the basis of invariant rules, it would have to appear unpredictable, downright random, even to its builders.81
Since modern ANNs implement a variant of such “learning,” Turing’s reply seems more plausible today than it did in the days of the sequential paradigm. Such a stance would reduce human authorship to the act of curating the output, which plays a role in all writing—at the very least in the speech act of the “imprimatur.” In the context of automated processes, one might locate it in the feedback process in which the programmer changes a code’s parameters based on the output produced. The less control that remains, the greater the role of editorial choice, one might say. [End Page 326] It is therefore plausible to think of authorship as editorship, as was already suggested during the heyday of hypertext literature.82
Going further, literary scholars have for a long time argued that the figure of the author may not be the most important thing when discussing a text. Rather, the meaning-making part of engaging with texts is the act of reading, not the act of writing. There is little difference between intentionally written and non-intentionally generated texts, after all, if claims to meaning can only be plausibly made on the side of the receiver, and this is true for AI writing as well.83 The role of the reader, the lector in fabula, would then deprivilege the author to such a degree that the actual production techniques of synthetic writing would become utterly unimportant. Such a blunt transfer of the discourse of authorial death to AI-generated texts would, however, diminish the possible contribution of literary studies.
For apart from the lector in fabula or reader-response criticism, a number of classical concepts could find a second use in the discussion of AI authorship, if they are sensibly updated. Julia Kristeva’s notion of “intertextuality,” for example, would not only make sense as the “mosaic of citations” of texts produced via primary authorship, but would invite its application to the medial situation of secondary, tertiary, and quaternary writing.84 Since the intertext here is quite literally present in the data-set—though not as an atomistic mosaic but as a holistic statistical distribution85—we are dealing with a multiplication of authorship relations.
Apart from the fact that programming languages themselves have authors, the very fabric of code is intertextually authored, since its functionality is highly recurrent, and programmers reuse both their own and others’ code snippets. This can happen via repositories like GitHub or forums like Stack Overflow, where either entire libraries of functions can be downloaded and plugged into own’s own code in a modular fashion, or where code sequences are collectively corrected or improved on a micro level.86 It is not uncommon here for authorship to be deliberately omitted. Mark Marino compares such code snippets to the use of a screw in mechanical engineering—an engineer would not cite its inventor every time the screw is used.87 The idea of collective authorship, which dominated research on net literature in the 1990s,88 is therefore also applicable to executable primary text, that is, program codes.
While this still refers to the interaction of human authors, science and technology studies go further and also include non-human actors in the writing collective, arguing for what one can call distributed authorship. Katherine Hayles, for example, explicitly assumes “multiple authorships” on the part of the computer’s hardware89 and suggests that calculating machines themselves can be thought of as “cognizers.”90 This stance is articulated primarily in the context of actor-network theory, which understands agency precisely not as “limited a priori to what ‘intentional,’ ‘meaningful’ humans do,”91 but as distributed across an assemblage of actors in complex chains of operations. [End Page 327] Instead of thinking of authors as solely human, in an ANT-inspired theory of authorship they would become a process and “the empirical person [would be] just one human actor in an actor-network called ‘author.’”92
Distributed authorship including the situated intertext, material resources, labor, and economic and legal relations (figure by Author).
This actor-network must be imagined as sprawling and immense. It would include not only the human author, now severely sidelined, and the intertext, now localized in the training set. Nor would it stop at the labor of those who wrote the training data or with the programmers responsible for the algorithms involved—either abstractly, as theorems (fig. 2 shows the formula for empirical risk minimization, a principle in statistical learning theory), or concretely, as code—as the “screws” in Marino’s metaphor. Moreover, the companies behind large language models (be they OpenAI, as fig. 2 suggests, or any other future major player), with their integration into the market and their extractive as well as normative impact on writers everywhere, are as much part of this network as the legal framework, which is still in the process of catching up with the realities of modern text AI.93 Finally, the network would even include all the matter involved: the machines that run the code as well as the minerals and rare earths they are made of, which are extracted from the soil in a process that [End Page 328] emits CO2, a direct factor in climate change. From this vantage point, the Earth itself would lay claim to be a participant in the authorship network.94
So conceived, distributed authorship as actor-network would be potentially infinite. Drawing the line between what still is and what is no longer part of it would be a pragmatic decision based on the questions one wants to ask of the network. And as relevant as distributed authorship is for checking the anthropocentric and falsely teleological tendencies of causal authorship, it is not without its problems either. Oddly enough, it is its very omnivorousness that may blur what is contained in its network. Large language models are a case in point: systems like ChatGPT derive their power from the size of the training data on which they are based, which now encompasses more or less the entire Internet. Thus, as these systems incorporate the language of the commons, “distributed authorship” would take on a new meaning that is more reminiscent of a collective unconscious as a source of “automatic” writing.95
Jasmin Meerhoff points out that there are drawbacks to such metaphors of collectivity, for the “maximally distributed, dispersed, or completely atomized authorship” implied here may mention economic dependencies but sublates them into a collective whole that can no longer be addressed as an ethically responsible agent.96 Here, it is advisable to avoid talk of a machine agency and instead acknowledge the actual human authors behind the AI output: as laborers, but also as bearers of identities.97 This call echoes considerations from the sociology of art, which emphasize the infrastructural “support personnel” that operate behind the seemingly autonomous entities of the art system and make its activities possible in the first place.98 Here, the overly personalizing causal authorship can in turn be a check on the overly collectivizing distributed authorship. Perhaps it is best to think of both as two countervailing forces—with one model also contained in the other—that together constitute the field of tension that is authorship in the age of machine learning.
Conclusion
Looking at the limited literary experiments conducted with AI to date, one would expect to see an emphasis either on the collectivity of the commons of language, or on the collaboration with the machine, both of which would suggest an increasing dissolution of the author function. With a few exceptions,99 however, authors seem to continue to attach their names to works produced in collaboration with AI systems. A book like Amor Cringe by the American poet K Allado-McDowell appears under the name of the human actor of its author network,100 even though it was explicitly written in cooperation with GPT-3.
Clinging to the name of the author, however, does not necessarily imply a reactionary reactivation of outdated ideas of genius or a plea for the total aesthetic autonomy of the artist. For the poet Allison Parrish, understanding computers as tools is less an ontological question than a political task: “It will always be a mistake to attribute [End Page 329] volition to the computer and not to the people who programmed it because attribution of volition is removing personal responsibility: the algorithm did it, not me.”101 Reclaiming authorship can be brought into play precisely as a defense against both the phantasm of the technically optimized AI genius and the absolute atomization of authorship, which no longer has a place for political, economic, and ethical responsibility. Instead of collaboration, clearly hierarchized co-creation is preferred here.102
As mentioned above, the claim to a truly strong AI authorship—capable of making autonomous works of art beyond collaboration, co-creation, and distribution—is found primarily in the tech industry, whose representatives understand art as a benchmark or a catalyst for the development of an artificial consciousness.103 Here, ideas of authorship as genius survive much more unquestioned than in the fields of art and literature, which have been busy deconstructing such grand narratives for more than sixty years. But it is also clear that the ideology of the Romantic author need not be human; it is malleable enough that the idea of absolute originality, creating something novel out of thin air, can certainly be applied to AI models, as Carys J. Craig and Ian R. Kerr show in their legal discussion of AI copyright.104 And so it is not in the literary engagement with language models, but in the Valley, that the Romantic genius seems to live on in AI research, conflating questions of consciousness and intelligence with intentional authorship. Literature, it appears, continues to pursue the model of NLP—the deliberate generation of text that may or may not be used, claimed, and repurposed by a human author who attaches his or her name to the final product.
Finally, the more we are surrounded by AI-generated writing—what Matthew Kirschenbaum has called the “textpocalypse”105—the less purchase we may, in the long run, give to the idea of “artificial authors.” As much as the decentering of distributed authorship requires a recentering via causal authorship, and vice versa, then, it is also possible that the concept of authorship as such will apply to fewer and fewer kinds of texts. This can be thought of as similar to the authorless situation that Foucault speculated about: “One can imagine a culture where discourses would circulate and be received without the author-function ever appearing.”106 This is already partly a reality in our culture: in the case of automated information systems, interfaces, and other machine communications to which we are exposed on a daily basis, the question of authorship does not arise at all—one would not infer from the display on an ATM that the machine is a person, nor would one essentially agonize over the original author of its messages.107 The more machine-generated text proliferates, the more likely it is to be regarded as neither natural nor artificial in Bense’s sense, but as what I have elsewhere called “post-artificial”—as essentially authorless, so that only its content remains to be examined, not its personal origin to be investigated.108 Of course, this may apply to certain genres more than others, and perhaps not at all to some. But at least as a heuristic concept, post-artificiality is a useful boundary for what might be possible in the age of machine-generated texts. The question of [End Page 330] authorship, in a late victory for Barthes, Foucault, and others, would in such a case truly be lost in the “anonymity of a murmur.”109
hannes bajohr is an assistant professor in the Department of German at the University of California, Berkeley. His research focuses on theories of the digital, the history of philosophy, and political theory. He is the author of Schreibenlassen: Texte zur Literatur im Digitalen (2022) and Postartifizielle Texte: Lesen im Zeitalter künstlicher Intelligenz (forthcoming 2025).
Notes
Versions of this paper were presented at the University of Basel, Collegium Helveticum Zurich, ETH Zurich, and LMU Munich. I am grateful to Jules Pelta Feldman, Maria Eriksson, Ute Holl, Markus Krajewski, Felix Stalder, Nora Al-Badri, Adrian Notz, Juan S. Guse, two anonymous reviewers, and others who provided many helpful suggestions for its improvement.
1. C. P. Snow, The Two Cultures (Cambridge: Cambridge University Press, 2012).
2. See, for an overview, Philipp Löffler and Tim Sommer, “Introduction: Institutions of Authorship,” Authorship 11, no. 1 (2023): 1–5.
3. For the distinction between strong and weak AI, see John R. Searle, “Minds, Brains, and Programs,” Behavioral and Brain Sciences 3, no. 3 (1980): 417–57. For the distinction between strong and weak artistic AI, see Hannes Bajohr, “The Paradox of Anthroponormative Restriction: Artistic Artificial Intelligence and Literary Writing,” CounterText 8, no. 2 (2022): 262–82.
4. It is revealing how common the word “genius” is in the interviews with computer scientists collected in Arthur I. Miller, The Artist in the Machine: The World of AI-Powered Creativity (Cambridge, MA: MIT Press, 2019). Adrian Daub describes how the word “genius” forms a core tenet of current Silicon Valley ideology, Adrian Daub, What Tech Calls Thinking: An Inquiry into the Intellectual Bedrock of Silicon Valley (New York: Farrar, Straus, and Giroux, 2020), chap. 3.
5. “Google Artists and Machine Intelligence,” http://ami.withgoogle.com.
6. Christopher T. Funkhouser, Prehistoric Digital Poetry: An Archaeology of Forms, 1959–1995 (Tuscaloosa, Ala.: University of Alabama Press); Saskia Reither, Computerpoesie: Studien zur Modifikation poetischer Texte durch den Computer (Bielefeld: Transcript, 2003); Scott Rettberg, Electronic Literature (London: Polity, 2019).
7. Nils J. Nilsson, The Quest for Artificial Intelligence: A History of Ideas and Achievements (Cambridge: Cambridge University Press, 2010), 103–13.
8. Warren Weaver, “Translation,” [1949] in Readings in Machine Translation, ed. Sergei Nirenburg, H. L. Somers, and Yorick Wilks (Cambridge, MA: MIT Press, 2003), 13–17.
9. Michael D. Gordin, Scientific Babel: How Science Was Done Before and After Global English (Chicago: The University of Chicago Press, 2015), 213–40.
10. Alan M. Turing, “Computing Machinery and Intelligence,” Mind 59, no. 236 (1950): 433–60.
11. Jay David Bolter, “Artificial Intelligence,” in Writing Space. The Computer, Hypertext, and the History of Writing (Hillsdale, NJ: Erlbaum, 1991), 171–193, here 180.
12. Joseph Weizenbaum, “ELIZA: A Computer Program for the Study of Natural Language Communication Between Man and Machine,” Communications of the ACM 9, no. 1 (1966): 36–45.
13. Michel Foucault, “What Is an Author?,” in Aesthetics, Method, and Epistemology, ed. James D. Faubion (New York: New Press, 1998), 211, 210.
14. Outside of this mainstream, however, there have been experiments with narrative structures in AI research, such as James Meehan’s TAIL-SPIN or Terry Winograd’s SHRDLU. See Leah Henrickson, Reading Computer-Generated Texts (Cambridge: Cambridge University Press, 2021), 11–14.
15. See for an alternative view of the entanglement of AI and NLP Avery Slater, “Post-Automation Poetics; or, How Cold-War Computers Discovered Poetry,” American Literature 95, no. 2 (2023): 205–27, http://doi.org/10.1215/00029831-10575035.
16. Noah Wardrip-Fruin, “Christopher Strachey: The First Digital Artist?,” Grand Text Auto, 2009, https://grandtextauto.soe.ucsc.edu/2005/08/01/christopher-strachey-first-digital-artist/.
17. Christopher Strachey, “The ‘Thinking’ Machine,” Encounter, no. 3 (1954): 25–31, here 25–26.
18. Strachey, “The ‘Thinking’ Machine,” 31.
19. Max Bense, “Über natürliche und künstliche Poesie,” in Theorie der Texte: Eine Einführung in neuere Auffassungen und Methoden (Cologne: Kiepenheuer & Witsch, 1962), 143–47. An English translation is available online: Max Bense, “On Natural and Artificial Poetry,” March 13, 2023, https://hannesbajohr.de/en/2023/03/13/max-bense-on-natural-and-artificial-poetry-1962/.
20. Max Bense, Ästhetik und Texttheorie, ed. Elisabeth Walther, vol. 3, Ausgewählte Schriften, (Stuttgart: Metzler, 1998), 385. Unless otherwise noted, all translations are mine.
21. Roland Barthes, “The Death of the Author,” in Image, Music, Text: Essays, trans. Stephen Heath (London: Fontana, 1977), 142–48. For the background of this well-known debate, see Seán Burke, The Death and Return of the Author: Criticism and Subjectivity in Barthes, Foucault and Derrida (Edinburgh: Edinburgh University Press, 2008).
22. Theo Lutz, “Stochastische Texte,” augenblick 4, no. 1 (1959): 3–9.
23. Toni Bernhart, “Beiwerk als Werk: Stochastische Texte von Theo Lutz,” editio, no. 34 (2020): 180–206, here 194.
24. Theo Lutz, “Über ein Programm zur Erzeugung stochastisch-logistischer Texte,” Grundlagenstudien aus Kybernetik und Geisteswissenschaft 1, no. 1 (1960): 11–16, here 11, 14; similarly, see Bense, Ästhetik und Texttheorie, 3:384.
25. Gerhard Stickel, “‘Computerdichtung’: Zur Erzeugung von Texten mit Hilfe von datenverarbeitenden Anlagen,” Der Deutschunterricht 18, no. 2 (1966): 120–125, here 123. See also Gerhard Stickel, “Monte-Carlo-Texte: Automatische Manipulation von sprachlichen Einheiten,” in Kunst aus dem Computer, ed. William E. Simmat (Stuttgart: Nadolski, 1967), 53–57. Stickel writes about the motivation and technical background of the autopoeme: “These computer-generated texts were created in late 1965, early 1966 on a playful whim. Like other colleagues at what was then the German Computing Center in Darmstadt, I was tired of writing service programs for various dictionaries just before Christmas. Why shouldn’t we use the computer to produce Christmas poems for once? I designed a simple slot-and-fill syntax, compiled semantically poor function words such as articles, pronouns, haben, sein, etc., and asked colleagues to suggest nouns, adjectives, full verbs, and adverbs on Christmas themes. I wrote a FORTRAN program for this dictionary, whose subroutines in FAP machine code also included a pseudo-random number generator. The number of lines (between six and twenty, because that’s how many still fit on a fanfold sheet), the sentence patterns (as a sequence of syntactically permitted slots), and the words filling the individual slots were chosen at random. There was no semantic control. It was only loosely indicated by the dictionary. The dictionary, syntax, and program were entered using punch cards. The program was very fast (when it finally ran) and produced three to four text-like structures per second. Only printing took considerably longer. The texts we liked were then selected from the thick fanfold stack. In other words, there were two human parts to the text genesis: the dictionary and the selection from the generated text set. . . . I ended up in the role of an experimental poet, which I didn’t really want at all. I then quickly wrote two short essays to counter the naïve concern that computers would take over the work of human poets in the foreseeable future.” Email to the author, April 8, 2021.
26. Jörgen Schäfer, “Passing the Calvino Test? Writing Machines and Literary Ghosts,” in Digital Media and Textuality: From Creation to Archiving, ed. Daniela Côrtes Maduro (Bielefeld: Transcript, 2017), 29.
27. Jean Baudot, La machine à écrire (Montreal: Editions du Jour, 1964).
28. Marc Adrian, Die Maschinentexte: Montagen, Textsynthesen, computergenerierte Texte und Permutationen. Sammlung 1966–1992, ed. Gerald Ganglbauer (Stattegg: Gangan, 2020); see for Adrian’s description of his own relationship to the Vienna group Marc Adrian, Inventionen (Linz: edition neue texte, 1980), 5.
29. Hannah B. Higgins, “An Introduction to Alison Knowles’s The House of Dust,” in Mainframe Experimentalism: Early Computing and the Foundations of the Digital Arts, ed. Hannah B. Higgins and Douglas Kahn (Berkeley: University of California Press, 2012), 195–99.
30. Dick Higgins, Computers for the Arts (Somerville: Abyss, 1970), 1.
31. Manfred Krause and Götz F. Schaudt, eds., Computer-Lyrik (Düsseldorf: Droste, 1968), 8.
32. Karl Krolow, “Die Seele wimmert,” Die Tat, January 13, 1968, 33.
33. Italo Calvino, “Cybernetics and Ghosts,” in The Uses of Literature, trans. Patrick Creagh (San Diego: Harcourt Brace Jovanovich, 1986), 3–27.
34. Hans Magnus Enzensberger, Einladung zu einem Poesie-Automaten (Frankfurt am Main: Suhrkamp, 2000), 52. Despite the late publication date, Enzensberger insists that his reflections on computer literature were completed in 1974.
35. Stefan Rieger, “‘Bin doch keine Maschine . . .’. Zur Kulturgeschichte eines Topos,” in Machine Learning: Medien, Infrastrukturen und Technologien der Künstlichen Intelligenz, ed. Christoph Engemann and Andreas Sudmann (Bielefeld: Transcript, 2018), 117–42; Philipp Schönthaler, Die Automatisierung des Schreibens und Gegenprogramme der Literatur (Berlin: Matthes & Seitz, 2022).
36. Franz Josef Czernin and Ferdinand Schmatz, “Notes about the Poetry Program POE,” Ars Electronica Archive, 1990, https://webarchive.ars.electronica.art/en/archives/festival_archive/festival_catalogs/festival_artikel.asp%3FiProjectID=8950.html; see also Peter Weibel, “Algorithmus und Kreativität: Algorithmendesign in der Literatur,” in Woher kommt das Neue? Kreativität in Wissenschaft und Kunst, ed. Walter Berka and Christian Smekal (Vienna: Böhlau, 2003), 85–97.
37. Roland Pohl, “Kann künstliche Intelligenz Lyrik? Nein, meint der Dichter Czernin,” Der Standard, February 23, 2023, https://www.derstandard.at/story/2000143824279/autor-czernin-ein-kreatives-programm-muesste-absichten-haben.
38. Racter, The Policeman’s Beard Is Half Constructed (New York: Warner, 1984).
39. Leah Henrickson, “Constructing the Other Half of The Policeman’s Beard, ” Electronic Book Review, April 4, 2021, http://doi.org/10.7273/2bt7-pw23.
40. Jorge Luis Borges, “The Garden of Forking Paths,” in Ficciones, ed. Anthony Kerrigan, trans. Helen Temple and Ruthven Todd (New York: Grove Press, 1962), 89–101; Nick Montfort, Twisty Little Passages: An Approach to Interactive Fiction (Cambridge, MA: MIT Press, 2003).
41. Gilles Deleuze and Félix Guattari, A Thousand Plateaus, trans. Brian Massumi (Minneapolis: University of Minnesota Press, 1987), 3–25.
42. Barthes, “The Death of the Author”; for a discussion of digital authorship in a specifically German context, see Florian Hartling, Der digitale Autor: Autorschaft im Zeitalter des Internets (Bielefeld: Transcript, 2009), 97–109.
43. George P. Landow, Hypertext: The Convergence of Contemporary Critical Theory and Technology (Baltimore: Johns Hopkins University Press, 1992), 14.
44. Carolyn Guertin, “Wanderlust: The Kinesthetic Browser in Cyberfeminist Space,” Extensions: Online Journal of Embodiment and Technology, no. 3 (2007), https://web.archive.org/web/20150921150355/http://www.performancestudies.ucla.edu/extensionsjournal/guertin.htm; see also Jasmin Meerhoff, “Verteilung und Zerstäubung: Zur Autorschaft computergestützter Literatur,” in Digitale Literatur II, ed. Hannes Bajohr and Annette Gilbert (Munich: edition text+kritik, 2021), 49–61.
45. That the very notion of authorship, in fact, did not disappear is discussed in Robert Coover, “Goldene Zeitalter: Vergangenheit und Zukunft des literarischen Wortes in den digitalen Medien,” Text+Kritik, no. 152 (2001): 22–30.
46. Christiane Heibach, “Ins Universum der digitalen Literatur: Versuch einer Typologie,” Text +Kritik, no. 152 (2001): 31–42, here 32.
47. Roberto Simanowski, “Autorschaften in digitalen Medien: Eine Einleitung,” Text+Kritik, no. 152 (2001): 3–21.
48. Christiane Heibach, Literatur im elektronischen Raum (Frankfurt am Main: Suhrkamp, 2003), 168–71.
49. Espen J. Aarseth, Cybertext: Perspectives on Ergodic Literature (Baltimore: The Johns Hopkins University Press, 1997), 132–36. For Aarseth, cyborg authorship comes in a variety of aspects: 1) “pre-processing” means preparing the program, loading it with input, and then letting the computer produce the text; 2) “co-processing” means using the computer during the text production; 3) “post-processing” means selecting from computer-produced text. These steps are ideal-types and can occur in various combinations. Aarseth appropriated the term from David Wall, “The World-Wide Web as Cyborg Author,” 1994, http://www2.iath.virginia.edu/courses/encr481/wall.paper.html. Recently, Scott Rettberg has applied it to ChatGPT, see Scott Rettberg, “Cyborg Authorship: Writing with AI – Part 1: The Trouble(s) with ChatGPT,” Electronic Book Review, June 2, 2023, http://doi.org/10.7273/5SY5-RX37.
50. Mattis Kuhn, Selbstgespräche mit einer KI (n.p.: 0x0a, 2021); Jörg Piringer, Datenpoesie (Klagenfurt: Ritter, 2018); Jörg Piringer, Günstige Intelligenz (Klagenfurt: Ritter, 2022); Hannes Bajohr, Halbzeug: Textverarbeitung (Berlin: Suhrkamp, 2018); 0x0a, ed., Poetisch denken, 4 vols. (n.p.: 0x0a, 2020).
51. Hannes Bajohr, Schreibenlassen: Texte zur Literatur im Digitalen (Berlin: August Verlag, 2022); Leah Henrickson, “Tool vs. Agent: Attributing Agency to Natural Language Generation Systems,” Digital Creativity 29, no. 2–3 (2018): 182–90.
52. Melanie Mitchell, Artificial Intelligence: A Guide for Thinking Humans (New York: Farrar, Straus, and Giroux, 2019).
53. Ethem Alpaydın, Machine Learning: Revised and Updated Edition (Cambridge, MA: MIT Press, 2021), 60–62.
54. The breakthrough came with David E. Rumelhart, James L. McClelland, and PDP Research Group, Parallel Distributed Processing: Explorations in the Microstructure of Cognition, Foundations, vol. 1 (Cambridge, MA: MIT Press, 1986); see Nilsson, Quest for Artificial Intelligence, 339–40.
55. Alpaydın, Machine Learning, 105–42; on the difference between sequential/connectionist, see Hannes Bajohr, “Algorithmic Empathy: Toward a Critique of Aesthetic AI,” Configurations 30, no. 2 (2022): 203–31.
56. Rishi Bommasani et al., “On the Opportunities and Risks of Foundation Models,” arXiv, 2021, http://arxiv.org/abs/2108.07258.
57. Funkhouser reports of a 2006 project to generate Rumi verses. Funkhouser, New Directions in Digital Poetry, 306–7.
58. Sepp Hochreiter and Jürgen Schmidhuber, “Long Short-Term Memory,” Neural Computation 9, no. 8 (1997): 1735–80.
59. Ashish Vaswani et al., “Attention Is All You Need,” Advances in Neural Information Processing Systems 30 (2017): 5999–6009; Long Ouyang et al., “Training Language Models to Follow Instructions with Human Feedback,” arXiv, 2022, http://arXiv.org/abs/2203.02155; for a comprehensive introduction to the Transformer architecture as it is used in ChatGPT, see Stephen Wolfram, “What Is ChatGPT Doing . . . and Why Does It Work?,” Stephen Wolfram Writings, February 14, 2023, https://writings.stephenwolfram.com/2023/02/what-is-chatgpt-doing-and-why-does-it-work.
60. Matthew G. Kirschenbaum, Track Changes: A Literary History of Word Processing (Cambridge, MA: Belknap Press, 2016); Martin Stingelin, “UNSER SCHREIBZEUG ARBEITET MIT AN UNSEREN GEDANKEN: Die poetologische Reflexion der Schreibwerkzeuge bei Georg Christoph Lichtenberg und Friedrich Nietzsche,” in Schreiben als Kulturtechnik: Grundlagentexte, ed. Sandro Zanetti (Berlin: Suhrkamp, 2012), 83–104. My point here is that repeating the truism of media studies and “Schreibszenenforschung” that “Our writing tools are also working on our thoughts”—and thus declaring writing with large language models essentially the same as writing with a typewriter—obscures what is genuinely new about writing with AI. For a position that heeds this insight and operates within the tradition of the “scene of writing,” see Moritz Hiller, “Es gibt keine Sprachmodelle,” in Noten zum “Schreiben,” ed. Davide Giuriato, Claas Morgenroth, and Sandro Zanetti (Paderborn: Fink, 2023), 279–85.
61. Florian Cramer, Exe.cut[up]able statements: Poetische Kalküle und Phantasmen des selbstausführenden Texts (Munich: Fink, 2011); for an overview of generative and combinatory techniques, see Scott Rettberg, “Combinatory Poetics,” chap. 2 in Electronic Literature (London: Polity, 2019).
62. For the performativity of code, see Hannes Bajohr, “Operative Ekphrasis: The Collapse of the Text/Image Distinction in Multimodal AI,” Word & Image, forthcoming.
63. François Chollet, Deep Learning with Python, 2nd ed. (Shelter Island: Manning Publications, 2021), 4.
64. Tom B. Brown et al., “Language Models Are Few-Shot Learners,” arXiv, May 28, 2020, http://arxiv.org/abs/2005.14165; OpenAI, “GPT-4 Technical Report,” 2023, http://doi.org/10.48550/ARXIV.2303.08774; James Manyika, “An Overview of Bard: An Early Experiment with Generative AI,” 2023, https://ai.google/static/documents/google-about-bard.pdf.
65. With LLaMA, Bloom, Mistral, and Falcon, open-source language models are available that indeed are open to be finetuned, even on a local computer. But while this means that state-of-the-art language models can be finetuned, it will most likely not do away with the widespread use of the “factory settings,” if only because the convenience of the default has always been a factor in the use of tools and services.
66. Brown et al., “Language Models Are Few-Shot Learners,” 48–49.
67. For this term, see Hannes Bajohr, “Dumb Meaning: Machine Learning and Artificial Semantics,” IMAGE 37, no. 1 (2023): 58–70.
68. The economic aspect of language models in particular and AI technology in general cannot be considered further at this point, but I want to note that there is ample literature on the subject: Dieuwertje Luitse and Wiebke Denkena, “The Great Transformer: Examining the Role of Large Language Models in the Political Economy of AI,” Big Data & Society 8, no. 2 (2021): 205395172110477; Kate Crawford, Atlas of AI: Power, Politics, and the Planetary Costs of Artificial Intelligence (New Haven: Yale University Press, 2021); Matteo Pasquinelli, The Eye of the Master: A Social History of Artificial Intelligence (London: Verso, 2023). That “ordinary” text processors, too, are entangled in economic structures is a point made by Till A. Heilmann, Textverarbeitung: Eine Mediengeschichte des Computers als Schreibmaschine (Bielefeld: Transcript, 2012); and Kirschenbaum, Track Changes.
69. K Allado-McDowell, Pharmako-AI (London: Ignota, 2020); Hannes Bajohr, (Berlin, Miami) (Berlin: Rohstoff, 2023). For the latter, I was still able to select the training set, unlike in the case of Allado-McDowell. I finetuned the open-source GPT-J and GPT-NeoX models on four contemporary German novels to derive a similar tone and subject matter to the input. For more information, see Hannes Bajohr, “Cohesion without Coherence: Artificial Intelligence and Narrative Form,” Transit, forthcoming.
70. See for instance Steve Tomasula, “Our Tools Make Us (and Our Literature) Post,” in The Bloomsbury Handbook of Electronic Literature, ed. Joseph Tabbi (London: Bloomsbury, 2018), 39–58.
71. Peter Dittmer, Schalten und Walten: Die Amme (Nuremberg: Verlag für moderne Kunst, 2006).
72. Ulf Stolterfoth, Ammengespräche (Schupfart: Urs Engeler, 2010).
73. Hannes Bajohr, “The Gestalt of AI: Beyond the Atomism-Holism Divide,” Interface Critique 3 (2021): 13–35. It is important to note that of course neural networks still run on digital systems and are thus only conceptually, but not computationally, distinct from older types of digital architectures. Likewise, neural nets can produce identical output if their “temperature” (their induced randomness) were reduced to zero; but this would go against their intended design.
74. Fabian Offert, “Can We Read Neural Networks? Epistemic Implications of Two Historical Computer Science Papers,” American Literature 95, no. 2 (2023): 423–28.
75. Leah Henrickson and Albert Meroño-Peñuela, “The Hermeneutics of Computer-Generated Texts,” Configurations 30, no. 2 (2022): 115–39.
76. Henrickson, “Tool vs. Agent”; Emanuele Arielli, “AI-Aesthetics and the Artificial Author,” Proceedings of the European Society for Aesthetics 15 (2023): 40–54. Note that the criterion for authorship here is weak—the unpredictability of the outcome—rather than strong—such as intention or social responsibility. See for a discussion of a “conservative” model of authorship that operates with strong criteria René Van Woudenberg, Chris Ranalli, and Daniel Bracker, “Authorship and ChatGPT: A Conservative View,” Philosophy & Technology 37, no. 1 (2024): 34, http://doi.org/10.1007/s13347-024-00715-1.
77. Joseph Weizenbaum, “Kunst und Computer,” in Computermacht und Gesellschaft: Freie Reden, ed. Gunna Wendt and Franz Klug (Frankfurt am Main: Suhrkamp, 2001), 103.
78. Turing, “Computing Machinery and Intelligence,” 450–51.
79. Ada Lovelace, “Translator’s Notes to M. Menebrea’s Memoir,” in Babbage’s Calculating Engines: Being a Collection of Papers Relating to Them; Their History, and Construction, ed. Henry P. Babbage (Cambridge: Cambridge University Press, 2010), 44. John Haugeland points out that Lady Lovelace’s objection follows a genealogical logic he finds questionable: “Why should an entity’s potential for inventiveness be determined by its ancestry (like some hereditary title) and not by its own manifest competence?” John Haugeland, Artificial Intelligence: The Very Idea (Cambridge, MA: MIT Press. 1985), 10. This would mean to cut the thread of causality espoused here and simply to look at the capabilities of the machine itself. The problem, as I see it, is that this argument locates some general “art-making capacity” solely within either the properties of the object created or the mechanism responsible for the creation, but ignores aspects like the historical and social circumstances of the work’s reception and its place within a community of judgment. See Hannes Bajohr, “The Deixis of Literature: On the Conditions of Recognizing Computers as Authors,” Orbis Litterarum, forthcoming.
80. I call this “Esprit der Regel”, Hannes Bajohr, “Vom Geist und den Maschinen: Autorschaft zwischen Mensch und Computer,” in Schreibenlassen, 33–38, here 37.
81. Turing, “Computing Machinery and Intelligence,” 459.
82. Uwe Wirth, “Der Tod des Autors als Geburt des Editors,” Text+Kritik, no. 152 (2001): 54–64.
83. Matthew G. Kirschenbaum, “Spec Acts: Reading Form in Recurrent Neural Networks,” ELH 88, no. 2 (2021): 361–86; Stephanie Catani, “‘Erzählmodus an’: Literatur und Autorschaft im Zeitalter künstlicher Intelligenz,” in Jahrbuch der deutschen Schillergesellschaft, vol. 64 (Berlin: de Gruyter, 2020), 287–310, here 304.
84. Julia Kristeva, Sèméiotikè: Recherches pour une sémanalyse (Paris: Seuil, 1969), 85.
85. This distinction lies, I believe, at the heart of old and new AI technology, see Bajohr, “The Gestalt of AI.”
86. Karl Wolfgang Flender, “Do Conceptualists Dream of Electric Sheep? Algorithmische Interpretation des Unbewussten in Conceptual Writing und konzeptueller Codeliteratur,” in Bajohr and Gilbert, Digitale Literatur II, 134–44, here 139.
87. Mark C. Marino, Critical Code Studies (Cambridge, MA: MIT Press, 2020), 50.
88. Heibach, Literatur im elektronischen Raum, 205–6.
89. N. Katherine Hayles, My Mother Was a Computer (Chicago: The University of Chicago Press, 2005), 105.
90. N. Katherine Hayles, “Can Computers Create Meanings? A Cyber/Bio/Semiotic Perspective,” Critical Inquiry 46, no. 1 (2019): 32–55. For an alternative view, see Bajohr, “Dumb Meaning.”
91. Bruno Latour, Reassembling the Social: An Introduction to Actor-Network-Theory, Clarendon Lectures in Management Studies (Oxford: Oxford University Press, 2005), 81.
92. Jens Schröter, “Autorschaft aus dem Blickwinkel der Akteur-Netzwerk-Theorie,” in Grundthemen der Literaturwissenschaft: Autorschaft, ed. Michael Wetzel (Berlin: De Gruyter, 2022), 626.
93. Kate Crawford and Vladan Joler give a much more exhaustive and impressive visualization of the actor-network of AI systems—not under the auspices of authorship but using Amazon’s Echo as an example. Crawford and Joler, “Anatomy of an AI System,” 2018, http://www.anatomyof.ai.
94. See Crawford, Atlas of AI; Emily M. Bender et al., “On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?,” in FAccT ’21: Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency (New York: Association for Computing Machinery, 2021), 610–23.
95. Meghan O’Gieblyn, “Babel: Could a Machine Have an Unconscious?,” n+1, July 27, 2021, https://www.nplusonemag.com/issue-40/essays/babel-4; Slavoj Žižek, “ChatGPT sagt das, was unser Unbewusstes radikal verdrängt,” Berliner Zeitung, April 7, 2024 (an English translation is available as “ChatGPT Says What Our Unconscious Radically Represses,” Sublation Magazine, April 7, 2023, https://web.archive.org/web/20240208020539/https://www.sublationmag.com/post/chatgpt-says-what-our-unconscious-radically-represses).
96. Meerhoff, “Verteilung und Zerstäubung,” 55.
97. This case is also made from the perspective of a practitioner of electronic literature by Jörg Piringer, “Elektrobarden,” Transistor 1, no. 2 (2019): 78–83.
98. Howard S. Becker, Art Worlds (Berkeley: University of California Press, 1982), x; Annette Gilbert, “Kollaterales Schreiben: Digitale Kollaboration im Zeitalter von Crowdworking und Algotaylorismus,” in Bajohr and Gilbert, Digitale Literatur II, 66.
99. See the examples in Manuel Kaufmann, Dreaming Data: Aspekte der Originalität und Autorschaft in der künstlichen Kreativität (Zurich: Chronos, 2022) and Jenifer Becker, “Können Chatbots Romane schreiben? Der Einfluss von KI auf kreatives Schreiben und Erzählen: Eine literaturwissenschaftliche Bestandsaufnahme,” in KI:Text ed. Gerhard Schreiber and Lukas Ohly (Berlin: De Gruyter, 2024), 83–100, http://doi.org/10.1515/9783111351490-007.
100. K Allado-McDowell, Amor Cringe (New York: Deluge, 2022).
101. Quoted in Miller, Artist in the Machine, 223.
102. David Jhave Johnston, ReRites: Human + A.I. Poetry: Raw Output (Montreal: Anteism Books, 2019); see also Kaufmann, Dreaming Data, 38.
103. See Daub, What Tech Calls Thinking; as an example, see Marian Mazzone and Ahmed Elgammal, “Art, Creativity, and the Potential of Artificial Intelligence,” Arts 8, no. 1 (2019).
104. Carys J. Craig and Ian R. Kerr, “The Death of the AI Author,” Ottawa Law Review 52, no. 1 (2019): 31–86.
105. Matthew Kirschenbaum, “Prepare for the Textpocalypse,” The Atlantic, March 8, 2023, https://www.theatlantic.com/technology/archive/2023/03/ai-chatgpt-writing-language-models/673318/.
106. Michel Foucault, “Qu’est-ce qu’un auteur?,” in Dits et écrits: 1954–1988, ed. Daniel Defert, François Ewald, and Jacques Lagrange (Paris: Editions Gallimard, 1994), 811. My translation. This passage does not appear in the English translation of the text.
107. Elena Esposito, Artificial Communication: How Algorithms Produce Social Intelligence (Cambridge, MA: MIT Press, 2022).
108. Hannes Bajohr, “On Artificial and Post-artificial Texts: Machine Learning and the Reader’s Expectations of Literary and Non-literary Writing,” Poetics Today 45, no. 2 (2024): 331–61.
109. Foucault, “What Is an Author?,” 222.