2
REAL-TIME COMPOSITION AND NOTATION IN NETWORK MUSIC ENVIRONMENTS Georg Hajdu Hochschule für Musik und Theater Hamburg www.georghajdu.de [email protected] ABSTRACT This paper presents an approach to real-time composition and notation implemented in the networked multimedia performance environment Quintet.net. The current version of the software uses MaxScore, a Max- Java object based on the Java Music Specification Language (JMSL). By utilizing another Max-Java object, the author exploits the tools to develop various scenarios (uni-, bi- and multi-directional), ultimately with the goal of involving classically trained musicians in the creative process. 1. INTRODUCTION Composition and notation of musical scores in real-time is one of the great remaining challenges of computer music, whose success story, particularly in recent years, involves the continuous refinement and dissemination of tools for sound synthesis and diffusion as well as gestural control of sound and multimedia. This has enabled electronic musicians to free their bodies from standard commercial controllers and, thereby, to create music performances in which gestures and motion are intricately mapped to various musical parameters in real-time [10]. In the meantime, spurred by the development of the Internet and its protocols, network music performance has become more and more popular. Performances which either focus on sending low-level sound over ever faster networks or setting up scenarios in which networked musicians manipulate higher-level processes via control messages [8]. Yet, network music performance is still too often the domain of expert musicians in front of an expert audience. What is generally lacking, some experiments by Gerhard Winkler [11] and others notwithstanding, is an involvement of classically trained musicians with their ability to (sight-) read music in standard music notation (SMN) [4]. While computer-assisted composition systems have been conceived and realized since the second half of the 1950s, most of them have either not been created with real-time application in mind, or lack the ability to instantaneously output the results in standard music notation. Sadly, environments that are capable of both require expert knowledge and/or haven’t reached the mainstream yet [12]. 2. MAXSCORE Since 2007, with the introduction of MaxScore [3], programmed by Nick Didkovsky—assisted by the author—, this gap is closing by offering the tools for Max/MSP (probably most popular computer music environment to this date; http://www.cycling74.com) to perform sophisticated music notation in real-time. This paper aims at demonstrating how this novel tool can be extended for use in network music performance. MaxScore is a Max-Java object, which is based on the Java Music Specification Language (JMSL) and its music notation package JScore [1][2]. It responds to more than 120 different messages. While the majority of these messages are being used to set up scores, create notes and manipulate their properties, a subset can be employed to query the properties of its elements (intervals, notes, tracks, staffs, measures, scores). Thus, MaxScore acts as a hierarchical matrix for standard music notation; it draws its graphics to the Max LCD- based canvas or bcanvas abstractions, much like a Jitter matrix draws to a jit.window or jit.pwindow. 3. QUINTET.NET Quintet.net [5] is a networked multimedia performance environment under continuous development by the author since 1999. The four components (Server, Client, Conductor and Viewer) are written in Max/MSP. Originally conceived for a quintet of musicians performing on the Internet by sending control messages to its built-in sampler, an open plug-in structure was implemented, allowing—thanks to the Max pattr system—the routing of an arbitrary input to an arbitrary output. From its inception, Quintet.net used music notation as a visual layer facilitating the interaction between the remote musicians by symbolically representing the music events created by them. A second notation layer permits the representation of scores and parts in SMN. The 2008 version integrates MaxScore but relies on another Java object (JScoreTranslator), written by Hungarian composer Ádám Siska, to translate MaxScore’s xml output into Max-compatible messages [9]. In addition, JScoreTranslator provides position and formatting information, which is being passed to Quintet.net’s built-in drawing engine. In the process, the amount of data is reduced to such an extent that the messages can easily be sent over WAN via OSC. Quintet.net was featured in numerous music events since its premiere in 2000, a Munich Biennale opera among them. In November 2007, Quintet.net was used simultaneously with CCRMA’s software JackTrip in a three-way performance of John Cage’s Five, connecting five musicians located in the cities of Belfast, Graz and Hamburg via a broadband multi-channel audio stream [7]. This was also the inaugural event of CO-ME-DI-A (Cooperation and Mediation in Digital Arts)—a Culture

Real Time Composition and Notation in Network Music

  • Upload
    ipires

  • View
    2

  • Download
    0

Embed Size (px)

DESCRIPTION

Real Time Composition and Notation in Network Music

Citation preview

  • REAL-TIME COMPOSITION AND NOTATION IN NETWORK MUSIC ENVIRONMENTS

    Georg Hajdu Hochschule fr Musik und Theater Hamburg

    www.georghajdu.de [email protected]

    ABSTRACT

    This paper presents an approach to real-time composition and notation implemented in the networked multimedia performance environment Quintet.net. The current version of the software uses MaxScore, a Max-Java object based on the Java Music Specification Language (JMSL). By utilizing another Max-Java object, the author exploits the tools to develop various scenarios (uni-, bi- and multi-directional), ultimately with the goal of involving classically trained musicians in the creative process.

    1. INTRODUCTION

    Composition and notation of musical scores in real-time is one of the great remaining challenges of computer music, whose success story, particularly in recent years, involves the continuous refinement and dissemination of tools for sound synthesis and diffusion as well as gestural control of sound and multimedia. This has enabled electronic musicians to free their bodies from standard commercial controllers and, thereby, to create music performances in which gestures and motion are intricately mapped to various musical parameters in real-time [10]. In the meantime, spurred by the development of the Internet and its protocols, network music performance has become more and more popular. Performances which either focus on sending low-level sound over ever faster networks or setting up scenarios in which networked musicians manipulate higher-level processes via control messages [8]. Yet, network music performance is still too often the domain of expert musicians in front of an expert audience. What is generally lacking, some experiments by Gerhard Winkler [11] and others notwithstanding, is an involvement of classically trained musicians with their ability to (sight-) read music in standard music notation (SMN) [4]. While computer-assisted composition systems have been conceived and realized since the second half of the 1950s, most of them have either not been created with real-time application in mind, or lack the ability to instantaneously output the results in standard music notation. Sadly, environments that are capable of both require expert knowledge and/or havent reached the mainstream yet [12].

    2. MAXSCORE

    Since 2007, with the introduction of MaxScore [3], programmed by Nick Didkovskyassisted by the author, this gap is closing by offering the tools for

    Max/MSP (probably most popular computer music environment to this date; http://www.cycling74.com) to perform sophisticated music notation in real-time. This paper aims at demonstrating how this novel tool can be extended for use in network music performance. MaxScore is a Max-Java object, which is based on the Java Music Specification Language (JMSL) and its music notation package JScore [1][2]. It responds to more than 120 different messages. While the majority of these messages are being used to set up scores, create notes and manipulate their properties, a subset can be employed to query the properties of its elements (intervals, notes, tracks, staffs, measures, scores). Thus, MaxScore acts as a hierarchical matrix for standard music notation; it draws its graphics to the Max LCD-based canvas or bcanvas abstractions, much like a Jitter matrix draws to a jit.window or jit.pwindow.

    3. QUINTET.NET

    Quintet.net [5] is a networked multimedia performance environment under continuous development by the author since 1999. The four components (Server, Client, Conductor and Viewer) are written in Max/MSP. Originally conceived for a quintet of musicians performing on the Internet by sending control messages to its built-in sampler, an open plug-in structure was implemented, allowingthanks to the Max pattr systemthe routing of an arbitrary input to an arbitrary output. From its inception, Quintet.net used music notation as a visual layer facilitating the interaction between the remote musicians by symbolically representing the music events created by them. A second notation layer permits the representation of scores and parts in SMN. The 2008 version integrates MaxScore but relies on another Java object (JScoreTranslator), written by Hungarian composer dm Siska, to translate MaxScores xml output into Max-compatible messages [9]. In addition, JScoreTranslator provides position and formatting information, which is being passed to Quintet.nets built-in drawing engine. In the process, the amount of data is reduced to such an extent that the messages can easily be sent over WAN via OSC. Quintet.net was featured in numerous music events since its premiere in 2000, a Munich Biennale opera among them. In November 2007, Quintet.net was used simultaneously with CCRMAs software JackTrip in a three-way performance of John Cages Five, connecting five musicians located in the cities of Belfast, Graz and Hamburg via a broadband multi-channel audio stream [7]. This was also the inaugural event of CO-ME-DI-A (Cooperation and Mediation in Digital Arts)a Culture

  • 2007 multi-annual collaboration project under the auspices of IRCAM (http://www.comedia.eu.org/).

    Figure 1. Data flow between Quintet.net components and program modules (the data formats are printed in italics).

    4. SCENARIOS FOR REAL-TIME COMPOSITION AND NOTATION

    Three types of scenarios can be distinguished for networked real-time composition and notation: A uni-directional one in which music is created by generative process and sight-read by the performers, a bi-directional one, in which the input from the musicians drives the generation of parts, and a related, multi-directional one, in which several sources participate in the generation of notated music in a more complex routing scheme. An example for the uni-directional scenario is the authors piece Ivresse 84 (2007) in which a stochastic process under the control of a conductor re-assembles a score (Cages 1st Freemann etude) in real-time [6]. Nick Didkovskys bi-directional piece Zero Waste (2002) takes a more interactive approach: Two measures are initially being (sight-)read by a performer, his/her playing transcribed and subsequently re-presented to the performer in a potentially endless feedback loop.

    This year, the author will be starting work on a composition in whichin a multi-directional scenariothe input from an ensemble of musicians will be used to create the piece in real-time in a branching process.

    5. CONCLUSION

    If the exploration of real-time composition and notation was ever impeded by the lack of appropriate tools, MaxScore now offers appropriate utilities for MaxMSP, one of the most common music programming environments. Integrating MaxScore into Quintet.net, leverages its capabilities for network performance. Three scenarios for music created in real-time by the machine and sight-read by capable musicians have been mentionedscenarios ranging from generative processes to interactive feedback loops involving one or several performers. Yet, it takes an experienced

    composers ingenuity to create a piece, which both takes advantage of the vast possibilities and yields results that are equally satisfying aesthetically.

    6. REFERENCES

    [1] Didkovsky, N. and Burk, P.L., (2001). "Java Music Specification Language, an introduction and overview." Proceedings of the International Computer Music Conference, pp. 123-126.

    [2] Didkovsky, N. (2004). "Java Music Specification Language, v103 update." Proceedings of the International Computer Music Conference, pp. 742-745.

    [3] Didkovsky, N. and Hajdu, G. (2008). "MaxScore, standard music notation in Max/MSP." Proceedings of the International Computer Music Conference.

    [4] Freeman, J. (2008). Extreme Sight-reading, Mediated Expression, and Audience Participation: Real-time Music Notation In Live Performance. Manuscript.

    [5] Hajdu, G. (2005). Quintet.net: An environment for composing and performing music on the Internet. Leonardo Journal 38(1).

    [6] Hajdu, G. (2007). Playing Performers. Proceedings of the Music in the Global Village Conference, Budapest, Hungary, pp. 41-42.

    [7] Hajdu, G. (2008). Getrennte Krper - Musiker im Netz. Positionen 74/2008, pp. 29-31.

    [8] Renaud, A. and Rebelo, P. (2006). Network Performance: Strategies and Applications. Proceedings of the 6th international conference on New Interfaces for Musical Expression

    [9] Siska, A. (2007). Pipelining between JMSL and Quintet.net. Proceedings of the Music in the Global Village Conference, Budapest, pp. 28-30.

    [10] Wessel, D. and Wright, M. (2002). Problems and Prospects for Intimate Musical Control of Computers Computer Music Journal (Volume 26.3, Fall 2002), pp. 11-22.

    [11] Winkler, G. (2004). The realtime-score. A missing link in computer-music performance. SMC04 Conference Proceedings.

    [12] Wulfson, H., Barrett, G. D. and Winter, M. (2007) Automatic Notation Generators Proceedings of the 7th international conference on New interfaces for musical expression, pp. 346 351.

    IndexICMC 2008 HomeConference InfoWelcome from the ICMA PresidentICMA OfficersWelcome from the ICMC 2008 Organising CommitteeICMC 2008Previous ICMCsICMC 2008 Paper Panel & Music CuratorsICMC 2008 ReviewersICMC 2008 Best Paper Award

    SessionsMonday, 25 August 2008Languages and Environments 1Interaction and Improvisation 1Sound SynthesisComputational Modeling of MusicDemos 1Posters 1Interaction and Improvisation 2Aesthetics, History, and Philosophy 1

    Tuesday, 26 August 2008MiscellaneousAlgorithmic Composition Tools 1Network PerformanceComputational Music Analysis 1Panel 1: Reinventing Audio and Music Computation fo ...Panel 2: Towards an Interchange Format for Spatial ...

    Wednesday, 27 August 2008Studio Reports 1Mobile Computer Ensemble PlayDemos 2Posters 2Algorithmic Composition Tools 2Interface, Gesture, and Control 1

    Thursday, 28 August 2008Interface, Gesture, and Control 2Languages and Environments 2Spatialization 1Computational Music Analysis 2Panel 3: Network PerformanceDemos 3Posters 3

    Friday, 29 August 2008Sound ProcessingAesthetics, History, and Philosophy 2Interface, Gesture, and Control 3Spatialization 2Algorithmic Composition Tools 3Studio Reports 2

    AuthorsAll authorsABCDEFGHIJKLMNOPQRSTUVWYZ

    PapersAll papersPapers by Sessions

    Topicscritical theory/philosophy of technology, postmodern cy ...sociology/anthropology of everyday sounds, situated per ...history of computer music, women and gender studies, ed ...philosophy/culture/psychology, music information retrie ...electroacoustic music composition, aesthetics of music, ...singing analysis/synthesis, music analysis/synthesis, v ...interactive and real-time systems and languages, music ...human-computer interaction, sound synthesis/analysis, i ...interaction design, computer music, performance art, el ...physical interface design, performance systems, gesture ...language/education/history/sociology of computer music, ...composition systems and techniques, languages for compu ...programming languages/systems, audio synthesis/analysis ...composition, music cognition, music informatics, human- ...music information retrieval, audio signal processing, p ...computational musicology, music cognition, music and AI ...music cognition, rhythm/meter/timing/tempo, computation ...music information retrieval, audio content analysis, to ...spatial audio, audio signal processing, auditory percep ...physical modelling, spatial audio, room acoustics, aura ...sonic interaction design, physics-based sound synthesis ...audio signal processing, sound synthesis, acoustics of ...audio signal processing, acoustics, software systemsphysics-based sound synthesis, virtual room acousticscomposition, music analysis, software for pedagogyPANEL: Towards an Interchange Format for Spatial audio ...PANEL: Network PerformancePANEL: Reinventing Audio and Music Computation for Many ...

    SearchHelpBrowsing the Conference ContentThe Search FunctionalityAcrobat Query LanguageUsing the Acrobat ReaderConfiguration and Limitations

    AboutCurrent paperPresentation sessionAbstractAuthorsGeorg Hajdu