{"id":309,"date":"2024-10-03T21:34:33","date_gmt":"2024-10-03T19:34:33","guid":{"rendered":"https:\/\/hannesbajohr.de\/en\/?p=309"},"modified":"2025-05-11T01:04:25","modified_gmt":"2025-05-10T23:04:25","slug":"rhetoric-as-codec-machine-learning-and-the-facilitation-of-reading-and-writing-with-an-aside-about-surfing-and-riding","status":"publish","type":"post","link":"https:\/\/hannesbajohr.de\/en\/2024\/10\/03\/rhetoric-as-codec-machine-learning-and-the-facilitation-of-reading-and-writing-with-an-aside-about-surfing-and-riding\/","title":{"rendered":"Rhetoric as Codec: Machine Learning and the Facilitation of Reading and Writing (With an Aside About Surfing and Riding)"},"content":{"rendered":"\n<p><em>This is a translation of a short intervention I was invited to contribute to Zeitschrift f\u00fcr Medienwissenschaft. You can read the original German text <a href=\"https:\/\/mediarep.org\/entities\/article\/68535382-8779-4c75-9688-a8b44ebe1870\" target=\"_blank\" rel=\"noreferrer noopener\">here<\/a>.<\/em><\/p>\n\n\n\n<p>To discuss large language models (LLMs) in the &#8220;Tools&#8221; section of the <em>Zeitschrift f\u00fcr Medienwissenschaft<\/em> is already to take a stand. After all, the debate about <a rel=\"noreferrer noopener\" href=\"https:\/\/www.tandfonline.com\/doi\/full\/10.1080\/14626268.2018.1482924\" target=\"_blank\">whether LLMs are tools or agents is far from settled<\/a>. The latter position does not even have to mean the fantasies of artificial general intelligence that OpenAI boss Sam Altman posits as the ultimate goal of any technical development. It might be enough to regard LLMs as partners in an &#8220;artificial communication&#8221; that are sufficiently unpredictable to create the &#8220;double contingency&#8221; of <a rel=\"noreferrer noopener\" href=\"https:\/\/mitpress.mit.edu\/9780262046664\/artificial-communication\/\" target=\"_blank\">communicative behavior.<\/a> And research on the <a rel=\"noreferrer noopener\" href=\"https:\/\/muse.jhu.edu\/article\/846197\" data-type=\"URL\" data-id=\"https:\/\/muse.jhu.edu\/article\/846197\" target=\"_blank\">&#8220;scene of writing&#8221; <\/a>(<em>Schreibszene<\/em>), abandoning the idea that texts are necessarily produced by humans, should also be able to warm to the reverse assumption that machines could be a &#8220;subject of writing&#8221; or participate in it.<span id='easy-footnote-1-309' class='easy-footnote-margin-adjust'><\/span><span class='easy-footnote'><a href='https:\/\/hannesbajohr.de\/en\/2024\/10\/03\/rhetoric-as-codec-machine-learning-and-the-facilitation-of-reading-and-writing-with-an-aside-about-surfing-and-riding\/#easy-footnote-bottom-1-309' title=' Moritz Hiller, \u201cEs gibt keine Sprachmodelle,\u201d in: Davide Giuriato, Claas Morgenroth, and Sandro Zanetti (eds.), &lt;em&gt;Noten zum \u201cSchreiben\u201d&lt;\/em&gt; (Paderborn: Fink, 2023), 280; see also Hannes Bajohr and Moritz Hiller, &lt;a rel=&quot;noreferrer noopener&quot; href=&quot;https:\/\/www.etk-muenchen.de\/search\/Details.aspx?fsize=1.1&amp;amp;SeriesID=0040-5329&amp;amp;ISBN=9783967079814&quot; target=&quot;_blank&quot;&gt;&lt;em&gt;Das Subjekt des Schreibens&lt;\/em&gt; &lt;\/a&gt;(Munich: edition text+kritik, 2024).'><sup>1<\/sup><\/a><\/span><\/p>\n\n\n\n<p>That I still treat large language models as tools is simply due to the fact that they are still primarily used as such at the moment. Two aspects of this tool-like nature deserve highlighting, even in the most abbreviated manner: how they facilitate reading and how they facilitate writing. <\/p>\n\n\n\n<p><\/p>\n\n\n\n<p><em>The Facilitation of Reading<\/em><\/p>\n\n\n\n<p>The first is located even before the text is produced, which is usually regarded as the actual product of the writing process. This is evident in translation services such as <a rel=\"noreferrer noopener\" href=\"https:\/\/deepl.com\/\" data-type=\"URL\" data-id=\"https:\/\/deepl.com\/\" target=\"_blank\">DeepL<\/a> or language transcription services like <a rel=\"noreferrer noopener\" href=\"https:\/\/otter.ai\/\" target=\"_blank\">Otter.ai<\/a>. An as yet much less discussed function seems to be the <em>facilitation of reading<\/em>, namely by summarizing what has already been written. If I ask Claude to reduce an essay to its basic arguments, I can get an idea of whether it is worth taking the time. Such &#8216;synoptability&#8217; seems to be closely tied to textual genres and disciplinary boundaries. Empirical social science articles, computer science white papers, even overlong encyclopedia entries can often be summarized sufficiently well. By contrast, a Lacan seminar, everyday communication based on things left unspoken as well as most kinds of poetry can rarely be reduced in a way that preserves what is not purely propositional to them. Those who write obscurely, one might say, will still not be read in the future \u2013 not only not by humans, but also not by machines. But this is hardly the fault of LLMs. Like all tools, they cannot be applied to just any domain, even if they, used as a hammer, make everything look like a nail.<\/p>\n\n\n\n<p>But where the digital humanities want to grapple with the great unread at a distance, the rest of the humanities may now indulge in the promise that machine learning will now also provide an overview at close range\u2014or, as Arno Schmidt once sighed, make the divergence between life-time and reading-time a little less steep.<span id='easy-footnote-2-309' class='easy-footnote-margin-adjust'><\/span><span class='easy-footnote'><a href='https:\/\/hannesbajohr.de\/en\/2024\/10\/03\/rhetoric-as-codec-machine-learning-and-the-facilitation-of-reading-and-writing-with-an-aside-about-surfing-and-riding\/#easy-footnote-bottom-2-309' title='Arno Schmidt, \u201cJulianische Tage,\u201d in: &lt;em&gt;Bargfelder Ausgabe&lt;\/em&gt;, Werkgruppe 3: &lt;em&gt;Essays und Biographisches&lt;\/em&gt;, vol 4: &lt;em&gt;Essays und Aufs\u00e4tze II&lt;\/em&gt; (Z\u00fcrich: Haffmanns, 1995), 87\u201392.'><sup>2<\/sup><\/a><\/span> All this means is that AI offers a solution to a problem for which it is itself responsible. If, as Matthew Kirschenbaum has stated, large language models cause a <a rel=\"noreferrer noopener\" href=\"https:\/\/www.theatlantic.com\/technology\/archive\/2023\/03\/ai-chatgpt-writing-language-models\/673318\/\" target=\"_blank\">&#8220;textpocalypse&#8221;<\/a> and flood the web and our lives with synthetic writing, then they also provide the means by which this flood can be controlled again: through summaries and textual condensations. <span id='easy-footnote-3-309' class='easy-footnote-margin-adjust'><\/span><span class='easy-footnote'><a href='https:\/\/hannesbajohr.de\/en\/2024\/10\/03\/rhetoric-as-codec-machine-learning-and-the-facilitation-of-reading-and-writing-with-an-aside-about-surfing-and-riding\/#easy-footnote-bottom-3-309' title='A &lt;a rel=&quot;noreferrer noopener&quot; href=&quot;https:\/\/x.com\/JeremyNguyenPhD\/-status\/1774021645709295840&quot; data-type=&quot;URL&quot; data-id=&quot;https:\/\/x.com\/JeremyNguyenPhD\/-status\/1774021645709295840&quot; target=&quot;_blank&quot;&gt;study&lt;\/a&gt; from earlier this year showed how realistic Kirschenbaum&amp;#8217;s prediction is: the phrase &amp;#8220;to delve into,&amp;#8221; one of ChatGPT&amp;#8217;s favorite phrases, now appears 10 to 100 times more often in PubMed articles than it did ten years ago.'><sup>3<\/sup><\/a><\/span><\/p>\n\n\n\n<p>This is remarkable in two ways. First, the summary as a means of reading becomes a means of writing again, namely when authors have their abstracts generated. Secondly, when I have a text summarized by an LLM that was already produced by an LLMs, the standard relationship between <em>compression<\/em> and <em>decompression<\/em> in information technology is reversed. Normally, the goal of information transmission is to keep the redundancy of a message as high as necessary in view of potential noise sources, but at the same time as low as possible in view of limited channel capacities. In the case of highly redundant natural language, compression is possible for transmission over a channel, followed by decompression at the receiver side (&gt;-&lt;). In the case of the &#8220;textpocalypse&#8221; mentioned above, however, the reverse happens: The message would go and come out compressed on both the sender and receiver sides, and the decompression would become the channel&#8217;s transmission codec (&lt;=&gt;).<\/p>\n\n\n\n<p>In this case, LLMs are more than just simple writing aids. Rather, as one could say with <a rel=\"noreferrer noopener\" href=\"https:\/\/cup.columbia.edu\/book\/man\/9780231052184\" target=\"_blank\">Arnold Gehlen<\/a>, Walter Benjamin, and <a rel=\"noreferrer noopener\" href=\"https:\/\/mitpress.mit.edu\/9780262521338\/work-on-myth\/\" target=\"_blank\">Hans Blumenberg<\/a>, they are tools for the psychosensory unburdening from the textual-absolute. There is evidence that that this observation is not entirely fanciful. Commercial programs such as the current Microsoft Office suite allow one to formulate entire e-mails from a series of keywords, while summarizing received messages in keyword form in turn. Elaboration is an interface between machines, not between humans, who receive only the reduced version of a text. Similarly, in the text-to-image AI DALL-E, the user&#8217;s input is no longer sent directly to the image generator, but is first embellished and provided with more details by the system. <a rel=\"noreferrer noopener\" href=\"https:\/\/cdn.openai.com\/papers\/dall-e-3.pdf\" target=\"_blank\">According to the OpenAI engineers<\/a>, such an <em>ornatus<\/em> leads to better results, but remains largely invisible to the user. This would reformulate the aper\u00e7u of programmer Andrej Karpathy, who emphasizes the new power of natural language (&#8220;<a rel=\"noreferrer noopener\" href=\"http:\/\/twitter.com\/karpathy\/status\/1617979122625712128\" target=\"_blank\">the hottest new programming language is English<\/a>&#8220;), thusly: &#8220;the hottest new transmission protocol is verbose English&#8221;. Rhetoric resides not only in the message, but also in the codec.<\/p>\n\n\n\n<p><\/p>\n\n\n\n<p><em>The Facilitation of Writing<\/em> <\/p>\n\n\n\n<p>Second, as mentioned before, the facilitation of reading already plays a role in the facilitation of writing. But AI-generated abstracts are just one example of those genres that could be called texts without <em>jouissance<\/em> and that seem to be especially suitable for AI writing assistance.<\/p>\n\n\n\n<p>For LLMs&#8217; big promise is this: to outsource the false writing time that is wasted on quality of life-spoiling routines such as proposals, administrative communication, and final reports to the machine in order to give research more real writing time. The fact that the DFG \u2013 the German NEH \u2013 now <a rel=\"noreferrer noopener\" href=\"https:\/\/www.dfg.de\/resource\/blob\/289674\/ff57cf46c5ca109cb18533b21fba49bd\/230921-stellungnahme-praesidium-ki-ai-data.pdf\" target=\"_blank\">explicitly allows this<\/a>, as long as the use of generative AI is identified, is surely not least due to the realization that more and more researchers&#8217; lives are squandered on an <em>\u0153uvre cach\u00e9e<\/em> of rejected proposals that no academic audience every gets to see.<\/p>\n\n\n\n<p>However, the distinction between false and real writing time may be misleading when &#8220;real&#8221; writing is also done without pleasure. It might therefore be more interesting to turn to literature as a presumed case of <em>jouissance<\/em>-filled text production. The writer <a rel=\"noreferrer noopener\" href=\"https:\/\/www.transcript-open.de\/doi\/10.14361\/9783839469224-013\" target=\"_blank\">Jenifer Becker described<\/a> her initial working mode with GPT-3 as a collective brainstorming, similar to the \u201cwriters&#8217; room\u201d in television series, in which ideas can be spitballed, taken up and discarded. Author Juan S. Guse recounts how he came to a similar assessment in two steps: While he initially used ChatGPT for a &#8220;collaboration <em>ex negativo<\/em>&#8221; in order to avoid &#8220;stochastically well-trodden paths&#8221;\u2014if the AI has the same thought as me, the thought is bad\u2014his practice has become more &#8220;maieutic&#8221; in that he now also accepts the program&#8217;s suggestions positively. <span id='easy-footnote-4-309' class='easy-footnote-margin-adjust'><\/span><span class='easy-footnote'><a href='https:\/\/hannesbajohr.de\/en\/2024\/10\/03\/rhetoric-as-codec-machine-learning-and-the-facilitation-of-reading-and-writing-with-an-aside-about-surfing-and-riding\/#easy-footnote-bottom-4-309' title='Juan S. Guse, &amp;#8220;Das kombinatorische Seekuh-Prinzip.&amp;#8221; In &lt;em&gt;Schreiben nach KI&lt;\/em&gt;, edited by Hannes Bajohr and Ann Cotten (Berlin: Rohstoff, forthcoming).'><sup>4<\/sup><\/a><\/span><\/p>\n\n\n\n<p>But these practices still seem to be about generating ideas, not literary texts. The latter is currently most likely to be produces in digital literature, which is particularly open to writing processes that exist at a certain distance from the author and at the same time do not have to follow conventional genre conventions.<span id='easy-footnote-5-309' class='easy-footnote-margin-adjust'><\/span><span class='easy-footnote'><a href='https:\/\/hannesbajohr.de\/en\/2024\/10\/03\/rhetoric-as-codec-machine-learning-and-the-facilitation-of-reading-and-writing-with-an-aside-about-surfing-and-riding\/#easy-footnote-bottom-5-309' title=' Simon Roloff and Hannes Bajohr. &lt;em&gt;Digitale Literatur zur Einf\u00fchrung&lt;\/em&gt; (Hamburg: Junius 2024; Hannes Bajohr, &amp;#8220;&lt;a rel=&quot;noreferrer noopener&quot; href=&quot;https:\/\/hannesbajohr.de\/en\/wp-content\/uploads\/sites\/2\/2023\/06\/Bajohr-Writing-at-a-Distance.pdf&quot; target=&quot;_blank&quot;&gt;Writing at a Distance: Notes on Authorship and Artificial Intelligence&lt;\/a&gt;.&amp;#8221; &lt;em&gt;German Studies Review&lt;\/em&gt; 47, no. 2 (2024): 315\u2013337.'><sup>5<\/sup><\/a><\/span>  My novel <em><a rel=\"noreferrer noopener\" href=\"https:\/\/www.matthes-seitz-berlin.de\/buch\/-berlin-miami.html?lid=2\" target=\"_blank\">(Berlin, Miami)<\/a><\/em>\u2014which I created using open LLMs called GPT-J and GPT-NeoX that I fine-tuned on contemporary literature\u2014also belongs to this genre.<span id='easy-footnote-6-309' class='easy-footnote-margin-adjust'><\/span><span class='easy-footnote'><a href='https:\/\/hannesbajohr.de\/en\/2024\/10\/03\/rhetoric-as-codec-machine-learning-and-the-facilitation-of-reading-and-writing-with-an-aside-about-surfing-and-riding\/#easy-footnote-bottom-6-309' title='An English-language excerpt appeared in &lt;em&gt;&lt;a rel=&quot;noreferrer noopener&quot; href=&quot;https:\/\/www.ensemblepark.com\/&quot; target=&quot;_blank&quot;&gt;Ensemble Park&lt;\/a&gt; &lt;\/em&gt;with an interview about the production in the print version.'><sup>6<\/sup><\/a><\/span> The text was not the result of a prompt (&#8220;Write me a novel!&#8221;), but came about through repeated sentence completion.<\/p>\n\n\n\n<figure class=\"wp-block-image alignleft size-large is-resized is-style-default\"><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/portal.dnb.de\/opac\/mvb\/cover?isbn=978-3-7518-7013-9\" alt=\"\" width=\"325\" height=\"531\"\/><figcaption class=\"wp-element-caption\">Hannes Bajohr, <em>(Berlin, Miami)<\/em>, Berlin: Rohstoff, 2023.<\/figcaption><\/figure>\n\n\n\n<p>In order to find out whether and in what way such a model is capable of narration, I intervened as little as possible and often simply let it run its course. The writing was therefore a particularly distanced affair. Nevertheless, I did not experience the LLM as an agent. Rather, I had the impression that it was the <em>text<\/em> that wanted something and was pushing in a certain direction. One is surfing on ideas, but they regularly end up absurd or digressive. For this reason, I actively intervened at some points, picking up lost leads or introducing new topics\u2014often with just one word\u2014that I wanted to know more about. For example, I came across the &#8220;Jawling&#8221;, invented by the model and only vaguely described, which, together with the equally vague &#8220;Pondhead&#8221;, was up to much mischief. Or I learned about the process of &#8220;diagonalization&#8221; to which the city of Miami, suffering from acute urban decay, was subjected under the aegis of the \u00c3\u00e4\u00e4 agency. <\/p>\n\n\n\n<p>This was all so interesting that I occasionally followed up when these topics threatened to disappear, but always remaining open to what else was there to come. The writing activity then consisted mainly of alternating between surfing and small nudges\u2014what in horseback riding is called &#8220;leg aids,&#8221; of which there are, among others, &#8220;leading&#8221; and &#8220;correcting&#8221; variants. <span id='easy-footnote-7-309' class='easy-footnote-margin-adjust'><\/span><span class='easy-footnote'><a href='https:\/\/hannesbajohr.de\/en\/2024\/10\/03\/rhetoric-as-codec-machine-learning-and-the-facilitation-of-reading-and-writing-with-an-aside-about-surfing-and-riding\/#easy-footnote-bottom-7-309' title=' I am not sure if I am translating this right. In German, the word is &amp;#8220;Schenkelhilfe&amp;#8221; and it exists in a &amp;#8220;vorw\u00e4rtstreibende,&amp;#8221; &amp;#8220;verwahrende,&amp;#8221; and even vorw\u00e4rts-seitw\u00e4rtstreibende variant, which might indicate an especially open-ended process. See Deutsche Reiterliche Vereinigung, ed. &lt;em&gt;Richtlinien f\u00fcr Reiten und Fahren&lt;\/em&gt;, vol. 1, &lt;em&gt;Grundausbildung f\u00fcr Reiter und Pferd&lt;\/em&gt; (Warendorf: Deutsche Reiterliche Vereinigung, 2014), 83. '><sup>7<\/sup><\/a><\/span> The experience of writing with AI moves between these metaphors: surfing and riding\u2014at least for me, at least for the moment. Here, the distinction between tool and agent becomes blurred. Just as a surfboard, no matter how much you experience it as an outgrowth of your own body, is not yet an agent, it would be wrong to speak of a horse as a tool. <span id='easy-footnote-8-309' class='easy-footnote-margin-adjust'><\/span><span class='easy-footnote'><a href='https:\/\/hannesbajohr.de\/en\/2024\/10\/03\/rhetoric-as-codec-machine-learning-and-the-facilitation-of-reading-and-writing-with-an-aside-about-surfing-and-riding\/#easy-footnote-bottom-8-309' title='At least this is how I imagine it; I can neither surf nor ride horses.'><sup>8<\/sup><\/a><\/span> In the end, LLMs may be a third thing for which we still have to find a practice and a name.<\/p>\n\n\n\n<hr class=\"wp-block-separator has-alpha-channel-opacity\"\/>\n\n\n\n<p><\/p>\n","protected":false},"excerpt":{"rendered":"<p>This is a translation of a short intervention I was invited to contribute to &#8220;Zeitschrift f\u00fcr Medienwissenschaft&#8221;.<\/p>\n","protected":false},"author":1,"featured_media":316,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[1],"tags":[6,10,14],"class_list":["post-309","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-allgemein","tag-ai","tag-digital-literature","tag-media-studies"],"_links":{"self":[{"href":"https:\/\/hannesbajohr.de\/en\/wp-json\/wp\/v2\/posts\/309","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/hannesbajohr.de\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/hannesbajohr.de\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/hannesbajohr.de\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/hannesbajohr.de\/en\/wp-json\/wp\/v2\/comments?post=309"}],"version-history":[{"count":12,"href":"https:\/\/hannesbajohr.de\/en\/wp-json\/wp\/v2\/posts\/309\/revisions"}],"predecessor-version":[{"id":381,"href":"https:\/\/hannesbajohr.de\/en\/wp-json\/wp\/v2\/posts\/309\/revisions\/381"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/hannesbajohr.de\/en\/wp-json\/wp\/v2\/media\/316"}],"wp:attachment":[{"href":"https:\/\/hannesbajohr.de\/en\/wp-json\/wp\/v2\/media?parent=309"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/hannesbajohr.de\/en\/wp-json\/wp\/v2\/categories?post=309"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/hannesbajohr.de\/en\/wp-json\/wp\/v2\/tags?post=309"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}