Live coding brings together music and code in live performance and has been practised for almost two decades [1]. Live coding using online crowdsourced sounds can be seen as a type of asynchronous collaboration. This approach can bring a rich, multifaceted range of timbres. However, there is a risk in the unpredictability of search results. The use of a virtual agent (VA) can complement a human live coder in her/his practice reducing this potential issue.
This live coding performance is a collaboration between a human live coder and a VA. MIRLCa1 is a self-built SuperCollider extension and a follow-up of the also self-built SuperCollider extension MIRLC [2]. The system combines machine learning algorithms with music information retrieval techniques to retrieve crowdsourced sounds from the online database Freesound [3], which results in a sound-based music style. In this performance, the live coder will explore the online database by only retrieving sounds predicted as “good” sounds when using the retrieval methods from the live coding system. This approach aims at facilitating serendipity instead of randomness in the retrieval of crowdsourced sounds.
A VA has been trained to learn from the musical preference of a live coder based on context-dependent decisions, ‘situated musical actions’ [4]. Adapted from Lucy Suchman’s concept of situated action [5], a situated musical action refers to any musical action related to a specific context, where the VA is expected to help the user in that action within that context. A binary classifier based on a multilayer perceptron (MLP) neural network has been used for sound prediction. For more background on aspects of the live coding tool used to develop this project, see [6][7][2].
This live coding performance is a collaboration between a human live coder and a virtual agent (VA). MIRLCa is a self-built SuperCollider extension and a follow-up of the also self-built SuperCollider extension MIRLC. The system combines machine learning algorithms with music information retrieval techniques to retrieve crowdsourced sounds from the online database Freesound.org, which results in a sound-based music style. In this performance, the live coder will explore the online database by only retrieving sounds predicted as “good” sounds when using the retrieval methods from the live coding system. This approach aims at facilitating serendipity instead of randomness in the retrieval of crowdsourced sounds. The VA has been trained to learn from the musical preference of a live coder within context-dependent decisions, ‘situated musical actions’. A binary classifier based on a multilayer perceptron (MLP) neural network has been used for sound prediction. The themes of legibility, agency and negotiability in performance will be sought through the collaboration between the human live coder, the virtual agent live coder and the audience. This project has been funded by the EPSRC HDI Network Plus Grant - Art, Music, and Culture theme.
The approximate duration of this performance is 15 minutes. It is also possible a shorter or longer duration. The preferred format is stereo online delivery, which can be either a real-time session or a pre-recorded session. The live coder will be sending a video stream showing her screen. Ideally, if there is an audience in a performance venue, there should be a projector showcasing the video stream and the in-house PA system can be used to amplify the audio sent from the video stream. The requirements are summarized next:
Stereo video stream sent by the live coder (alternatively a pre-recorded video).
If there is an audience in a performance venue:
In-house PA system.
A projector.
This solo performance is improvised. Previous online solo performances of related work have been presented at:
Transnodal TOPLAP, 2021.
IKLECTIK Offsite in London, 2020.
Sound Junction Satellites University of Sheffield Concerts/Algomech, 2020.
Network Music Festival, 2020.
Eulerroom Equinox, 2020.
This project has been funded by the EPSRC HDI Network Plus Grant - Art, Music, and Culture theme.