ABSTRACT

This chapter analyzes the process of remixing through the evolution of machine-driven text transformation approaches. It also analyzes the remixes of text's evolving structure since the early experiments in the 1960s, the role of machines in effectively emulating a writer's style and their essential support to generate credible fakes, particularly deepfakes, discussing a series of relevant issues for Digital Humanities. Fakes are usually produced to make a statement about their own content or form. Since the 1950s, early experiments with programming code to generate texts in various forms and styles were mostly centered on a combinatorial approach. And the outputted texts are equally remixed, based on some preexisting work and a varying number of database entries. There is ecology of literary bots online, especially Twitter bots, which are endlessly remixing content from databases or other online sources.