The Rutgers Camden Digital Studies Center
Facebook, November 16 - 21, 2016
A ntoinette LaFarge is an artist and writer working with impersonation, virtuality, and history as her main subjects. Her work takes form as computer-mediated performance, interactive installation, digital prints, and writing. Recent new media projects include Far-Flung follows function (2013), Galileo in America (2012), Hangmen Also Die (2010), and WISP (World-Integrated Social Proxy, 2009-10). She co-curated two early exhibitions on computer games and art: SHIFT-CTRL: Computers, Games, and Art (2000) and ALT+CTRL: A Festival of Independent and Alternative Games (2003), the latter supported by an NEA grant.
LaFarge's work has been seen internationally in venues from UpStage to documenta, and her writing has appeared in a range of journals and anthologies. She is currently engaged on a series of projects that spring out of her work as a Wikipedia editor. Recent writing about new media includes essays on avatarism, imposture as living fiction, and social proxies. More information is available on her webpage at http://www.antoinettelafarge.com
She is professor of digital media in the Art Department, Claire Trevor School of the Arts, UC Irvine.
Mixed Reality Performance
A ll of my mixed-reality performance works use different authoring strategies and tools. In general, however, all share a focus on multi-authoring, on improvisation in various forms, and on a fluid relationship between creation of text and creation of other forms, including software, vocals, sound, video, and movement.
Perhaps the best introduction to my authorial practices is a 2004/2006 mixed-reality performance work entitled "Demotic". The essential idea behind "Demotic" was to have a single stage actor and two sound artists channeling, in real time, many other voices and information sources, some from real space, and some from the Internet. I conceived it with director Robert Allen, and we co-created it with actor Tracey A. Leigh, sound artists Maria de los Angeles Esteves and Jeff Ridenour, and a group of online performers known as the Plaintext Players.
The following general description of our processes applies to both the 2004 and 2006 versions of the piece, but it should be noted that not all the methods described were used in every segment of the final work. To start with, we gathered the Plaintext Players on a MOO, which is a virtual, text-based, multi-user domain of a kind that predominated on the Net before the advent of graphical worlds. The MOO performers improvised 'in character', in real time, creating a text that was partly written, partly performed. Only one of the MOO performers was situated in the same physical space as the actor and the sound artists, and this person served as a key link between the remote and local performers. The base text generated by the MOO improvisers was fed into the physical space of the stage in two ways: as visuals and as sound. The visuals took the form of scrolling text projections, which the stage performer used as a kind of teleprompter, responding vocally (by reading/improvising) and physically (with improvised movement).
Since the MOO is a form of programmable software, we were able to control and alter the text output from the MOO: for example, in some cases we altered the speed with which it appeared on screen, or its layout, while in others we algorithmically garbled it or mixed the live text with previously recorded (stored) material.
The transformation of the MOO text into sound took two forms in addition to the actor's own reading of the text. One was artificial speech created by text-to-speech synthesis, which gave the remote MOO performers a vividly physical presence in the stage space. The other was a layered soundscape created by further processing of the MOO text and synthetic speech through the programming environment known as MAX/MSP, which allowed our sound designers to deploy spatialization, repetition, additional sounds, and other effects.
Feedback loops were a critical part of this 'interdependent' creative process. Through streaming audio, the remote MOO performers could hear what the stage actor and sound artists were doing with their text in real time (though slightly delayed) and respond to it. And since the sound designers and the actor were in the same physical space, they could respond directly to each other as well as working with the MOO text as it was created.
Although from a traditional writerly perspective, one could argue that 'the text' was created in the very first step (during the MOO improvisations), from our perspective what mattered was that it then underwent a series of transformations each of which brought new creative elements into play: at the point of MOO output, at the point of speech synthesis, at the point of sound processing, at the point of actor improvisation, at the point of feedback...Only at the end of this web of authorship did we have what we would consider 'the text'; that is, the full synthesis of verbal, audio, visual, and physical elements that is "Demotic."
For a diagram of the authorial process, see the "Demotic" website at:
Information about the specific processes featured in each segment of "Demotic"can be found in the program notes at:
The words above are also available on Antoinette LaFarge's content | code | process page. A transcript introduces her discussion on issues of trolling on social media and her words are also incorporated on other discussions throughout Social Media Narrative: Issues in Contemporary Practice.