A Novel Antenna Based Controller for Musical Expression
This paper introduces Tune Field, a 3-dimensional tangible interface that combines and alters previously existing concepts of topographical, field sensing and capacitive touch interfaces as a method for musical expression and sound visualization. Users are invited to create experimental sound textures while modifying the topography of antennas. The interface’s touch antennas are randomly located on a box promoting exploration and discovery of gesture-to-sound relationships. This way, the interface opens space to playfully producing sound and triggering visuals; thus, converting Tune Field into a sensorial experience.
Musical controller, Real-time expression, Shape-Changing Tangible Interface, Capacitive touch Interfaces, Sculptable Surface.
•Sound and music computing→Drone Synth; •Hardware → PCB design and layout; Tactile and interfaces; •General and reference → HCI and Design.
There are two ways to make a classical sculpture; by addition or subtraction, so it happens with music and sound waves. Tune Field is influenced by this idea in its physical interaction, through a series of antennas that slide along the Z axis, and the sound generated, a minimalist drone characterized by its slight harmonic variations[1].
Creating music required a basic knowledge to play the right note on an instrument or a knob on a synthesizer. In the 20th century, more experimental music-making interfaces have been created that incorporate randomness and allow the user to interact with non-instrumental objects to create sound. This concept has been explored in Tune Field using antennas as an interface.
Antennas are commonly used in radio engineering as transducers to convert the voltage from a transmitter into a radio signal. In Tune Field, we measure its capacitance, allowing us to recognize its position in the Z-axis. In other words, as the antenna’s length varies so does its capacitance. Furthermore, since antennas can pick radio signals out of the air and convert them into a voltage, the neighboring antennas will interfere with the measurement of each individual one. Therefore, the electric field generated will interfere in the sound produced converting the antenna-scape into a sound sculpture constantly altered.
Finally, the drone sound created is complemented by a series of visuals that create an experience inspired by the one proposed by La Monte Young in Dream House[2]. An installation comprised of static sounds and light sculptures.
Tune Field’s uniqueness comes from the fact that it is able to combine contactless signal modulation and capacitive sensing techniques. Beside the aesthetic and sculptural advantages, Tune Field provides a UI/performance advantage building on the physicality of a knob or a slider with the advances explored by contactless instruments.
In an effort to convey the concept of sculpting topography, two concepts emerged: the idea of stretching a fabric or creating a physical topography. In the first one, there are existing examples of instruments that use capacitive sensing fabrics, but this technology has limited application and relies on complex knitting techniques[3][4]. There are also some existing examples of topography using pin-based, actuated shape display technology like TRANSFORM[5]. Kinéphone builds on it to recreate instruments or serve as a music controller[6]. However, this solution relies on complex and expensive technology.
In the human-computer interaction field capacitive sensing has played a prominent role[7]. There are a lot of commercially available interfaces that can sense in the x/y axis, from the well-known Kaoss Pad[8] to the smartphone. Some of them also recognize the Z axis through pressure or doing a global, single z axis capacitive sensing. However, they cannot physically visualize it. As an alternative, antennas' capacitive touch capability represents a low-cost approach that can address this challenge.
The main instrument related to antennas is the Theremin. Although some consider it the greatest musical wonder of our time[9] its sound is limited to two axes, volume and pitch. However, some of the successors, such as the Theremini[10], defeats this limitation with its MIDI function. Years later, Max Mathews with its Radio Baton was able to compute the X, Y and Z coordinates of each Baton[11] These two innovations influenced many others in the field of contactless musical instruments and other applications based on electric field sensing[12], such as ScanFish[13] that was used in the brain opera[14].
Tune Field interface design consists of a series of antennas playfully distributed on a box with no labels that encourages spontaneous interactions. The lack of guidelines in the interface encourages users to touch, slide and change the position of the antennas to understand the impact on the sound and visuals produced.
Each antenna works as a capacitive touchpoint connected to one pin of Teensy 3.2 microcontroller. The PCB board design includes two Teensy 3.2 boards, one of them connected to a Teensy audio adaptor[15]. The initial design included the possibility of having an autonomous instrument where one of the microcontrollers analyzed the user interaction while the other generated the synthesis. It would be produced by the Teensy audio library as Grain Prism does[16]. The audio library[17], similar to Pure-Data or Max/MSP, is limited to the computational capability of the microcontroller, that is why we opted to produce its sound externally in Max/MSP.
Unlike other capacitive sensing interfaces, the antennas can recognize their height in the Z-axis and if it is being touched. These features allow replacing the use of other existing interfaces such as buttons or sliders. It also plays a collective role due to the magnetic field created which interferes in the capacitive sensing measurement of each antenna. This converts Tune Field into a novel interface that allows users to press, slide, tune or switch with a single interface touchpoint, an antenna.
The values collected from each antenna are processed through Teensy’s built-in Capacitive Sensor Library. Teensyduino sends the data by OSC first to Max/MSP to produce the sound, and later to Processing for the visuals, “Figure 4.”.
The antennas are sending continuous data which is averaged and smoothed before synthesis. When the antenna is touched there is a spike in the data that is discarded. The values recorded are scaled to later be used in the FM synthesis. Each antenna corresponds to a particular partial in the analysis frame[18]. On the other hand, analyses are made of the sound of Gong, Cello and Duduk samples with the Sigmund~ object[19]. In the original sound, we look at the 10 loudest partials and transfer these values forward to be combined with the input value of each antenna. Each partial from the analysis is resynthesized with that information using FM synthesis[20] creating a complex of new frequencies out of each analysis frequency. As a result, these sounds complexes come together to create a rich and evolving drone.
Once the sound is generated, it is further fed into a Sigmund~ object for analysis, and the 20 loudest frequency and amplitude pairs are collected and passed forward by OSC to the Processing application.
Processing, the programming language for creative coding, creates the visuals for the drone sound with two overlapping elements, a collection of waves and a series of color bars. The waves visualize the last 20 amplitudes of the sound generated. Each bar represents a particular partial of the sound. The height is mapped to the value of the signal’s frequency, in hertz, and the weight corresponds to the amplitude of that particular frequency, “Figure 5.”.
In the analysis frame, future work includes dedicating an antenna to select a specific instrument (Gong, Cello or Duduk). Also, this version uses static analysis, ideally, this would be continuous.
The visuals created are exclusively controlled by the sound spectrum. Future iterations will explore visualizations that rely less on color and more on shape or contrast as a visual differentiator.
Finally, the interface should explore different PCB designs and antenna arrangements, in a music sequencer, or by tones, frequencies etc. This would also impact how the antennas interfere with each other and modify the surrounding magnetic field.
This paper features Tune Field, an experimental and experiential musical interface composed of a series of antennas. This instrument allows users to interact and compose music in a novel way. Each antenna will see an impact on its height and on the overall magnetic field generated. Therefore, the sound and visuals produced can be altered converting this instrument into a ‘living’ sound sculpture that is constantly evolving.
https://www.media.mit.edu/projects/tune-field/overview/
I would like to thank Professors Matthew Schumaker, Mark Feldmeier, and Zach Lieberman, as well as Thomas Sanchez and Andres Rico for their help with this project.