Top Banner
Spectral Analysis, Editing, and Resynthesis: Methods and Applications Michael Kateley Klingbeil Submitted in partial fulfillment of the requirements for the degree of Doctor of Musical Arts in the Graduate School of Arts and Sciences COLUMBIA UNIVERSITY 2009
167
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Klingbeil Dissertation Web

Spectral Analysis, Editing, and Resynthesis:Methods and Applications

Michael Kateley Klingbeil

Submitted in partial fulfillment of therequirements for the degree

of Doctor of Musical Artsin the Graduate School of Arts and Sciences

COLUMBIA UNIVERSITY

2009

Page 2: Klingbeil Dissertation Web

© 2008Michael Kateley Klingbeil

All Rights Reserved

Page 3: Klingbeil Dissertation Web

ABSTRACT

Spectral Analysis, Editing, and Resynthesis: Methods and Applications

Michael Kateley Klingbeil

This document describes the design, development, and applications of cross-platform

software for audio analysis, editing, and resynthesis. Analysis is accomplished using a

variation of the McAulay-Quatieri technique of peak interpolation and partial tracking.

Linear prediction of the partial amplitudes and frequencies is used to determine the

best continuations for sinusoidal tracks. A high performance user interface supports

flexible selection and immediate manipulation of analysis data, cut and paste, and un-

limited undo/redo. Hundreds of simultaneous partials can be synthesized in real-time

and documents may contain thousands of individual partials dispersed in time without

degrading performance. A variety of standard file formats, including the Sound Descrip-

tion Interchange Format (SDIF), are supported for import and export of analysis data.

Specific musical and compositional applications, including those in works by the author,

are discussed.

Page 4: Klingbeil Dissertation Web

Contents

List of Figures iv

List of Tables vi

Acknowledgements vii

1 Introduction 1

1.1 Timbre and the Computer . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1

1.2 Synthesis Methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5

1.3 Theory Versus Implementation . . . . . . . . . . . . . . . . . . . . . . . . . . 9

1.3.1 Implementation History . . . . . . . . . . . . . . . . . . . . . . . . . . 10

1.3.2 Present Work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11

2 Spectral Modeling 13

2.1 The Frequency Domain . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13

2.1.1 Sinusoids . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13

2.1.2 The Fourier Transform . . . . . . . . . . . . . . . . . . . . . . . . . . . 14

2.1.3 The Discrete Fourier Transform . . . . . . . . . . . . . . . . . . . . . 16

2.1.4 The Short-Time Fourier Transform . . . . . . . . . . . . . . . . . . . . 18

2.2 STFT Analysis and the Phase Vocoder . . . . . . . . . . . . . . . . . . . . . . 22

2.2.1 Putting Phase to Good Use . . . . . . . . . . . . . . . . . . . . . . . . 24

2.2.2 Resynthesis and Transformation . . . . . . . . . . . . . . . . . . . . . 25

2.2.3 Time Expansion/Compression . . . . . . . . . . . . . . . . . . . . . . 26

2.2.4 Transposition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28

2.2.5 Refinements . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29

2.3 Spectral Envelopes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30

i

Page 5: Klingbeil Dissertation Web

2.4 Sinusoidal Modeling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31

2.4.1 STFT Frequency Interpolation . . . . . . . . . . . . . . . . . . . . . . 32

2.4.2 Partial Tracking . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35

2.4.3 Improved Partial Tracking . . . . . . . . . . . . . . . . . . . . . . . . . 37

2.4.4 Synthesis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40

2.4.5 Limitations and Extensions . . . . . . . . . . . . . . . . . . . . . . . . 41

3 Sinusoidal Analysis and Resynthesis 42

3.1 Analysis Parameters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 42

3.1.1 Analysis Window . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 42

3.1.2 Amplitude Thresholds . . . . . . . . . . . . . . . . . . . . . . . . . . . 44

3.2 Partial Tracking Using Linear Prediction . . . . . . . . . . . . . . . . . . . . 46

3.3 Resynthesis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48

3.3.1 Inverse FFT Synthesis . . . . . . . . . . . . . . . . . . . . . . . . . . . 48

3.3.2 IFFT Synthesis Implementation . . . . . . . . . . . . . . . . . . . . . . 52

3.3.3 Oscillator Bank Resynthesis . . . . . . . . . . . . . . . . . . . . . . . . 53

3.3.4 Phase Management . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54

3.4 Data Storage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56

3.4.1 Time-span Frames . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57

3.5 Transient Sharpening . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 58

3.6 Noise Modeling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61

3.6.1 Bandwidth Enhanced Sinusoids . . . . . . . . . . . . . . . . . . . . . 62

4 Editing and Transformation 64

4.1 Graphical User Interface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 64

4.2 Selection Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67

4.2.1 Rule-based Selection . . . . . . . . . . . . . . . . . . . . . . . . . . . . 68

4.3 Editing Features . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 68

4.4 Time Expansion and Contraction . . . . . . . . . . . . . . . . . . . . . . . . . 70

4.5 Frequency Modifications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 72

ii

Page 6: Klingbeil Dissertation Web

4.6 Data Exchange . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 72

5 Compositional Applications 78

5.1 Spectral Composition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 78

5.2 Sonic Transformations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 82

5.2.1 Timbre Hybridization . . . . . . . . . . . . . . . . . . . . . . . . . . . 82

5.2.2 Spectral Tuning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 84

5.3 Software Composition Environments . . . . . . . . . . . . . . . . . . . . . . 85

5.3.1 OpenMusic . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 86

5.3.2 Max/MSP . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 87

5.3.3 Common Music . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 90

5.4 Compositional Examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92

5.5 Performance Models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 95

6 Conclusions 97

References 101

A Tear of the Clouds score 107

A.1 Program Note . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 107

A.2 Score . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 108

iii

Page 7: Klingbeil Dissertation Web

List of Figures

2.1 DFT magnitude spectrum of a sinusoid with period N4 . . . . . . . . . . . . 18

2.2 DFT magnitude spectrum of a sinusoid with period N4.21 . . . . . . . . . . . 19

2.3 Sinusoids of 4.21 and 4.0 cycles per N samples . . . . . . . . . . . . . . . . . 19

2.4 DFT magnitude spectrum and the interpolated underlying magnitude spec-

trum of a sinusoid with period N4.21 . . . . . . . . . . . . . . . . . . . . . . . . 21

2.5 DFT magnitude spectrum and the interpolated underlying magnitude spec-

trum of a sinusoid with period N4 . . . . . . . . . . . . . . . . . . . . . . . . . 22

2.6 Rectangular window . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23

2.7 Triangular window . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23

2.8 Hann window . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23

2.9 Blackman window . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23

2.10 Overlap-adding the product of Blackman analysis and synthesis windows

with varying overlap factors . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28

2.11 Spectral envelope of a soprano voice . . . . . . . . . . . . . . . . . . . . . . . 30

2.12 Phase vocoder analysis of a frequency sweep . . . . . . . . . . . . . . . . . . 32

2.13 Sinusoidal modeling analysis of a frequency sweep . . . . . . . . . . . . . . 32

2.14 Estimating a peak in the spectrum based on a parabolic curve fitting . . . . 33

2.15 Connecting STFT peaks into sinusoidal tracks . . . . . . . . . . . . . . . . . 35

2.16 Partial tracking conflicts due to glissandi . . . . . . . . . . . . . . . . . . . . 37

3.1 Magnitude spectrum of the main lobes of two Blackman windows with

complete separation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 44

3.2 Magnitude spectrum of the main lobes of two Blackman windows with the

minimum allowable frequency separation . . . . . . . . . . . . . . . . . . . . 44

iv

Page 8: Klingbeil Dissertation Web

3.3 Frequency dependent threshold curve with the default values . . . . . . . . 46

3.4 Comparison of sinusoids generated via IFFT synthesis . . . . . . . . . . . . 52

3.5 Family of cubic phase interpolation functions with different values of M . . 55

3.6 Data structure for division of partials into time-span frames . . . . . . . . . 57

3.7 Breakpoints from the analysis of a castanet click using the standard model 60

3.8 Breakpoints from the analysis of a castanet click using time reassignment . 60

4.1 Zoomed-out display . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65

4.2 Zoomed-in display . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65

4.3 The SPEAR tool palette . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66

4.4 Playback control sliders . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66

4.5 Different breakpoint storage models . . . . . . . . . . . . . . . . . . . . . . . 69

4.6 Original partials prior to time expansion . . . . . . . . . . . . . . . . . . . . 71

4.7 Partials time expanded in independent mode . . . . . . . . . . . . . . . . . . 71

4.8 Partials time expanded in proportional mode . . . . . . . . . . . . . . . . . . 71

4.9 Dialog boxes for transposition and frequency shifting . . . . . . . . . . . . . 72

4.10 Frequency flipping . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73

4.11 The par-text-frame-format specification . . . . . . . . . . . . . . . . . . . . 76

4.12 Sample data as par-text-frame-format . . . . . . . . . . . . . . . . . . . . . 76

4.13 The par-text-partials-format specification . . . . . . . . . . . . . . . . . . 77

4.14 Sample data as par-text-partials-format . . . . . . . . . . . . . . . . . . . 77

5.1 OpenMusic patch . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 86

5.2 OpenMusic notation editor . . . . . . . . . . . . . . . . . . . . . . . . . . . . 87

5.3 OpenMusic patch that reads and writes SPEAR data . . . . . . . . . . . . . 88

5.4 Max/MSP patch from the CNMAT Spectral Synthesis Tutorials . . . . . . . 89

5.5 Dynamic retuning with crossing partials . . . . . . . . . . . . . . . . . . . . 95

v

Page 9: Klingbeil Dissertation Web

List of Tables

1.1 Digital synthesis techniques . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5

4.1 Some of the standard SDIF matrix types . . . . . . . . . . . . . . . . . . . . . 74

4.2 Supported file formats . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 75

5.1 Transformation functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91

5.2 Frequency adjustment functions . . . . . . . . . . . . . . . . . . . . . . . . . 92

5.3 Spectral envelope functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . 93

vi

Page 10: Klingbeil Dissertation Web

Acknowledgements

I wish to thank my advisor Tristan Murail and committee members Brad Garton and Fred

Lerdahl for their insight and support throughout my time at Columbia University (and

beyond). Thanks also goes to James Beauchamp for introducing me to the details of signal

processing, to Rick Taube for revealing the realm of the metalevel, and to Karl Klingbeil

for opening my ears to the world of contemporary music. Special gratitude goes to Anita

Buckmaster for her unwavering encouragement and assistance.

This dissertation is dedicated to my parents, Laura Kateley and Jack Klingbeil.

September 17, 2008

vii

Page 11: Klingbeil Dissertation Web

1

1. Introduction

1.1 Timbre and the Computer

Considerable energy and effort has been applied to further the understanding of timbre,

one of music’s most elusive features. The American National Standards Institute defines

timbre in a purely exclusionary manner: “. . . that attribute of sensation in terms of which

a listener can judge that two sounds having the same loudness and pitch are dissimilar.”

(ANSI S1.1-1960(R1976)-12.9) This rather unsatisfying description essentially relegates all

aspects of sound other than pitch and loudness to a single perceptual category. Since at

least the beginning of the twentieth-century, the conception of timbre has taken on an

increasingly important role in Western compositional practice. In 1922 Arnold Schoenberg

asserted, “The evaluation of tone color, the second dimension of tone, is in a much less

cultivated, much less organized state than is the aesthetic evaluation of pitch. Nevertheless,

we go right on boldly connecting sounds with one another, contrasting them with one

another, simply by feeling. . . ” (Schoenberg 1983). More than eight decades later, this

statement still seems regrettably accurate. Schoenberg goes on to call for a theory of timbre,

analogous to pitch theories, that will enable the composition of “tone color melodies.”

While falling far short of a unified theory, scientific and artistic efforts of recent decades

have achieved a greatly expanded understanding of timbre. For example, while Schoenberg

describes tone color as a singular “second dimension” of tone, there is now strong evidence

suggesting that timbre is in some sense multi-dimensional (Grey 1977). It is now theorized

that sensations of pitch and timbre are closely intertwined, and that we likely experience

different classes of pitch sensation (spectral pitches and virtual pitches) (Terhardt et al.

Page 12: Klingbeil Dissertation Web

1. Introduction 2

1982). What is undeniable, whether a composer grounds her work in theory, or spirituality,

or chooses to work “simply by feeling,” is that technological resources have had a profound

impact on recent musical practice, and that an understanding of timbre is essential to

effectively harnessing these resources.

The application of the digital computer, in particular, has done much to expand the

practical range of sonic possibilities and also to enable the theoretical and empirical study

of timbre. In a prophetic lecture given in 1936, Edgard Varèse proposed a vision of music

made possible by machine:

. . . the new musical apparatus I envisage, able to emit sounds of any numberof frequencies, will extend the limits of the lowest and highest registers . . .Not only will the harmonic possibilities of the overtones be revealed in alltheir splendor, but the use of certain interferences created by the partialswill represent an appreciable contribution. . . . The time will come whenthe composer, after he has graphically realized his score, will see this scoreautomatically put on a machine that will faithfully transmit the musical contentto the listener. (Varèse and Chou 1966, 12)

Varèse was hardly the first, nor last, to make such predictions, and yet the strength of

his conviction and the precision of his forecast remain striking. By 1953, when Varèse

would begin his first compositional experiments with magnetic tape, numerous composers,

technicians, and researchers were already at work in studios throughout the world ex-

ploring the musical applications of electronic instruments. By 1957, with the first musical

applications of general purpose digital computers at Bell Labs, the stage was set for the

eventual realization of Varèse’s predictions.

It was in fact the advent of digital sound synthesis that made this vision possible.

Computer music pioneer Max V. Matthews understood the power and generality of digital

synthesis which he described in his 1963 article in Science: “With the aid of suitable output

equipment, the numbers which a modern digital computer generates can be directly

converted to sound waves. The process is completely general, any perceivable sound can be so

produced.1” (Matthews 1963, 553). For the composer seeking to explore a world of timbral

possibilities, the notion is seductive. And yet the computer music composer, whether

1My emphasis

Page 13: Klingbeil Dissertation Web

1. Introduction 3

working in 1967 or forty years later, still faces a number of formidable challenges. We can

summarize these particular problems as follows: the time/cost problem, the synthesis

problem, the control problem, and the compositional problem. Let us briefly consider each

of these issues in turn.

Digital music production distinguished itself from all previous musical-mechanical

approaches in that sound creation could occur on a time scale completely distinct from the

temporal evolution of the sound itself. At first these computations were ponderous (many

orders of magnitude slower than real-time), but as computational power increased over the

decades, reasonably complex and credible sounds could be computed faster than real-time.

The implication of this temporal decoupling is that sound complexity is no longer bound

by space constraints (the number of musicians one can fit on a stage, or the number of

inputs on a mixing board, for example) but by limitations of time (and/or expense). The

composer now has the capability, as Varèse envisioned, to “emit sounds of any number

of frequencies,” and yet at the same time now grapples with a continual “compromise

between interest, cost, and work” (Matthews 1963, 555). One could argue that this problem

is being tackled with continual efficiency improvements in computer hardware. Compared

to the IBM 704 of 1957, the laptop computer of the early twenty-first century provides

a five-hundred-thousand fold increase in processing power at at fifteen-hundred times

less cost (Sandowsky 2001). Nevertheless, computer musicians seem to have a particular

penchant for using spare processing power.

With decreased costs and staggering increases in computing speed, significant addi-

tional challenges still remain. Although “any conceivable sound” is certainly possible, in

reality, the range of potential sounds is constrained by the development of appropriate

synthesis algorithms. The ability to formulate these procedures is severely limited by our

understanding of timbre and psychoacoustics (Matthews 1963). This can be classified as

“the synthesis problem.”

The control problem is a closely associated concern. Although digital synthesis “allows

composers freedom of expression at the timbral level” it also entails the “additional

burden of specifying the microstructure of sound, which was formerly the domain of the

Page 14: Klingbeil Dissertation Web

1. Introduction 4

performing musician” (Loy 1989, 326). Because interesting musical sounds are rarely static,

nor entirely stochastic, the precise control parameters must be specified at every instant.

If this control information is not a natural result of the synthesis procedure in use, the

result is an unwieldy data explosion with separate detailed control streams for each sonic

parameter of interest.

Finally, the composer is faced with all of the traditional problems of composition

itself — matters of intention and aesthetics, as well as possible questions of melody, coun-

terpoint, rhythm, harmony, and/or form.

The following list summarizes the challenges and some possible approaches:

Challenge Possible ApproachesI. Time and Cost faster computers, higher processing density, more

efficient algorithmsII. Synthesis problem new synthesis methods, flexible software systems

III. Control problem new user interfaces (visual, gestural), computerassisted composition, machine learning/listening

IV. Compositional problem continued exploration, theory

It is furthermore recognized that these problems are inherently interrelated. A given

composition aesthetic, one that involves indeterminacy, for example, would invite a

particular approach to the specification of control information. Particularly with regard to

the compositional “problem,” it is certain that no ultimate “solution” exists. If it did, such

a discovery would spell the end of creative artistic production and individual expression.

The act of composition must be seen as a continual search, for which each artist may be

free to define their own goals and strategies.

The chapters that follow are concerned in large part with particular approaches to the

synthesis problem, with a discussion of the resulting implications for the cost and control

problems. The motivation will always be oriented towards musical goals, which will be

addressed in the penultimate chapter.

Page 15: Klingbeil Dissertation Web

1. Introduction 5

1.2 Synthesis Methods

Regardless of compositional approach the practitioner of computer music always faces

the same question: What sounds should I produce? Although current computer music

systems offer tremendous flexibility, the choices are constrained by current theories and

their implementations. In order to contextualize the work described in subsequent chapters,

it may prove instructive to consider a particular taxonomy of digital synthesis techniques

developed by Smith (1992). Table 1.1 maintains the original categories while adding some

more recent methods. A significant development in the past decade is the emergence of

hybrid approaches.

Time Domain Spectral Models Physical Models Abstract MethodsMethods (Frequency Domain)

Sampling Additive Waveguide FMConcrète Phase Vocoder Finite Element AMGranular Sinusoidal (MQ/PARSHL) (CORDIS-ANIMA) WaveshapingScanned Synthesis SMS (Smith & Serra) Karplus-Strong Analog simulation

Subtractive Modal (VCO, VCA, VCF)Source-filter (LPC) Chaotic systemsUPIC (Xenakis)FOF/FOG/PSOLA

WavetableParticles (Roads)

ConcatenativeAudio Mosaicing (MPEG-7)

Physically Informed Sonic Modeling (PhISM)

Table 1.1: Taxonomy of current digital synthesis techniques (after Smith 1992).

Let us first consider some of the most commonly used synthesis methods. Since the

early 1990s, sampling synthesis has been the predominant sound production method

used in commodity hardware and software. Sampling offers very high fidelity results

(limited only by the quality of the input), but is restricted in terms of expressive and

transformational possibilities. A typical sampling synthesis architecture limits transfor-

mations to amplitude envelopes, filtering, and loop control. Commercial developers are

attacking the limitations of sampling in a brute force manner. It is not uncommon to

find multi-gigabyte sample libraries containing complex mappings, cross-faded overlays,

Page 16: Klingbeil Dissertation Web

1. Introduction 6

and keyswitch programs. For example, a complex violin bank might include samples

for legato, marcato, spiccato, staccato, and tremolo bowings all recorded at a variety of

different dynamic levels. Fast secondary storage allows these massive sample banks to be

streamed from disk in real-time.

Granular synthesis may be viewed as a particular specialization of sampling synthesis.

The granular approach overlaps many short sound segments (generally less than 50 msec.

duration) which may be derived from recorded sources or produced via other synthetic

means. Usually, each sound segment is modulated by a corresponding amplitude envelope

that quickly fades-in, reaches a steady state, and then fades out. Granular synthesis offers

the possibility to decouple the time and frequency evolution of sound, as well as impart

particular characteristics modulating between rough, smooth, or stochastic textures.

Wavetable synthesis, one of the oldest techniques, can be viewed as a type of hybrid

approach. The time domain view concentrates on the particular wave shape of the table,

which may be either derived mathematically, drawn interactively, or (most commonly)

extracted from a recorded sample. In the frequency domain perspective, the contents of

the wavetable are specified in terms of the amplitude, frequency, and phase of a set of

harmonically related partials. With both approaches, wavetable interpolation is necessary

to produce lively and dynamic sounds.

Particle synthesis may be viewed as a hybrid extension of granular synthesis that

attempts to build up specific frequency domain content using a variety of elementary sound

types which might include grains, impulses (trainlet or pulsar synthesis), or chirps (glisson

synthesis). Convolution may be used to impart specific frequency domain characteristics

to the “pulslet” texture. For a more detailed discussion of these methods, consult Roads’s

Microsound (2001, ch. 4).

Frequency domain models are concerned first and foremost with the spectral content

of sounds. This is a particularly powerful technique because the perception of timbre

is closely related to spectral content. Auditory research shows that the ear behaves as a

sort of spectrum analyzer with precise physical positions along the basilar membrane of

the inner ear attuned to specific frequencies. Frequency domain methods may thus be

Page 17: Klingbeil Dissertation Web

1. Introduction 7

seen as way to model synthesis on the processes of audition. Frequency domain models

usually follow an analysis/synthesis paradigm in which control parameters are derived by

analyzing pre-existing recordings.

Additive synthesis is one of the oldest and still most flexible of these frequency domain

methods. Oscillators tuned to specific frequencies, each with time varying amplitudes, are

mixed (summed) together to form a composite timbre. Typically, each oscillator produces

a sine-like waveform. In some applications the frequency and/or phase2 of each oscillator

may also be time variant. While theoretically capable of generating almost any type of

sound, additive synthesis is particularly unattractive in light of the control problem. A

complex timbre may easily require 100 individual oscillators that, in turn, each require

their own amplitude and frequency functions. This problem is typically solved by deriving

synthesis data from an analysis stage. The major advantage of additive synthesis, and

frequency domain methods in general, is that they are highly amenable to transformations.

The phase vocoder has been particularly favored for its support for independent time

and frequency modification. Andy Moorer’s Lions Are Growing, a work from 1978 that

time stretches, compresses, and transposes speech sounds,3 offers a clear demonstration

of the artistic potential of this approach. Beginning in the early 1990s, the technique

became possible on commodity computing platforms and received widespread use in

both commercial sound design and music composition. Spectral models, with particular

emphasis on sinusoidal models, will be taken up in further detail in subsequent chapters.

Concatenative synthesis, a category that encompasses a variety of specific techniques,

may be viewed as a hybrid of spectral and time domain methods. For example, attack

transients might be most effectively handled with sampling, while steady states could

be executed with a spectral model. By carefully concatenating4 sound segments derived

from a large database, more convincing and dynamic note-to-note transitions and timbral

modulations can be achieved than with sampling or spectral synthesis alone. A number of

2The phase of any non-zero frequency oscillator is by definition time varying — what is emphasized here isthat phase may be modulated directly rather than being a result of a fixed or time varying frequency.

3The source sound is a reading of Richard Brautigan’s poem of the same name. The work employsLPC-style resynthesis coupled with phase vocoding.

4Generally trying to preserve phase continuity

Page 18: Klingbeil Dissertation Web

1. Introduction 8

recently released commercial products such as Synful (orchestral instrument synthesis)

(Lindemann 2007) and Vocaloid (singing voice synthesis) (Bonada and Serra 2007) have

demonstrated the viability of this technique. A more detailed overview of concatenative

methods may be found in Schwarz (2005).

In the early 1990s, Julius O. Smith predicted “abstract-algorithm synthesis seems

destined to diminish in importance,” (Smith 1992) and although this proved true to some

extent, a resurgent interest in classic analog and digital hardware has demonstrated that

these methods retain a strong following. As a result, considerable efforts in both the

commercial and academic sectors have gone toward the software emulation of classical

voltage controlled oscillators and filters.5 In recent years the “vintage emulation” trend

has extended to early commercial digital synthesizers, including the venerable Yamaha DX

series which has been replicated in a number of software packages.6

It seems certain that no particular technique will ever disappear entirely from the

toolbox, rather the collection will expand with the exigencies of fashion dictating the

most popular methods of the time. The flexibility and modularity of current software

systems suggests that hybrid synthesis methods will become increasingly important. The

somewhat arbitrary boundaries between approaches will blur as practitioners explore

possible combinations and concatenations. After all, sampling, wavetable, granular, and

pulsar synthesis differ solely on their operative time scales (and approach to periodicity)

and thus could be subsumed under a single methodology.

Physical modeling represents a considerably different approach to sound production.

While sampling, spectral, and abstract models are all understood in terms of their effect

on the sound consumer (the ear), physical modeling is concerned with the details of the

sound producer (the instrument) (Smith 1992). The physics of musical instruments serve

as the starting point. Generally a set of differential equations are used to describe the

possible motion (vibrations) in response to physical input (such as the motion of a bow or

the stroke of a mallet). A virtual sound pressure wave can then be computed. A significant

5The interested reader might wish to consult Välimäki and Huovilainen (2006) as well as Stilson (2006).6Most notably with Native Instruments FM-7 series (now FM-8 in its 2007 incarnation)

Page 19: Klingbeil Dissertation Web

1. Introduction 9

benefit is that a very rich and sonically convincing output can result from simple control

inputs.

Despite many compelling advantages, physical modeling has yet to become perva-

sive in computer music composition. The lack of generality in many physical modeling

implementations may be one impediment to their acceptance. Even with a pre-designed

algorithm, the parameter space may be undesirably abstract and difficult to control. For

example, in a woodwind model the output pitch may be a result of multiple inputs. Partic-

ular physical modeling methods (for example waveguides) may be limited to modeling

certain restricted classes of instruments (for more detail see Smith 2004). Finite-difference

time-domain methods (Cadoz et al. 1993) represent a more general approach to physical

modeling, although the computational demands are still somewhat prohibitive for many

systems of musical interest. It is certain, however, that with continued research and

development and inevitable increases in processing speed, both this and other physical

modeling methods will find a following. After all, one must consider that even in the

relatively short history of computer music, these methods are still fairly new.

1.3 Theory Versus Implementation

Regardless of the technique of interest, the computer music practitioner is always limited

by the tools and implementations at hand. In some cases the constraints of the tools

may prove compositionally beneficial in their circumscription of the available musical

materials. At other times, the realities of the tools (missing features, bugs, awkward

interface, sluggish operation) may prove merely frustrating. Too often one finds a lack of

available implementations of some of the most compelling synthesis and signal processing

techniques described in the literature. In such situations computer music composers may

find it necessary to build their own tools — be it in software or hardware. Computer

music is an inherently multi-disciplinary endeavor involving researchers, mathematicians,

programmers, engineers, theoreticians, composers, improvisers, and performers, and in

most cases one individual may be closely tied to a number of these disciplines.

Page 20: Klingbeil Dissertation Web

1. Introduction 10

Over the past several decades, many such individuals have devoted considerable

attention to the area of spectral modeling. The proliferation of various spectral modeling

implementations attests to its appeal. Some of the technique’s advantages (noted above)

are especially attractive to composers — in particular the possibility to explore timbre

in a manner that is directly related to real-world experience (unlike abstract algorithm

synthesis). The transformative potential of spectral modeling synthesis (largely unavailable

with sampling) opens up a vast space between purely mimetic or purely synthetic sounds.

Sinusoidal modeling has a proven track record of high quality resynthesis while offering

numerous possibilities for novel sonic transformations (Pampin 2004). The technique

(which will be described in detail in section 2.4) derives a detailed additive synthesis

model from an input sampled sound signal. Something that closely resembles the original

input sound (a resynthesis) can be generated by adding together a sufficient number of

individual sinusoidal waves. In most cases the resynthesis will not be identical to the

original sound, although it is possible to get very close and, with certain extensions to the

basic technique, it is possible to recreate the original exactly.

1.3.1 Implementation History

A complete implementation of sinusoidal modeling was first detailed in paper by Robert J.

McAulay and Thomas F. Quatieri (1986), who were working at MIT on the analysis

and synthesis of speech signals. The basic technique is often referred to as the MQ

method. This work was closely followed by the development of the PARSHL system for

analysis-resynthesis of musical tones by Julius O. Smith and Xavier Serra (1987) at Stanford

University. Serra extended this into the “Spectral Modeling Synthesis” (SMS) method

which added residual noise modeling to the basic sinusoidal analysis (Serra 1989). At the

same time Robert C. Maher (1989), working with James Beauchamp at the University of

Illinois, was using sinusoidal modeling for polyphonic voice separation. Much of Maher’s

code was subsequently integrated into the SNDAN suite (Beauchamp 1993).

Page 21: Klingbeil Dissertation Web

1. Introduction 11

Maher’s work also served as the basis for Lemur by Kelly Fitz, Lippold Haken, and

Bryan Holloway (Fitz et al. 1995). Lemur, which ran on the Macintosh, was one of the

first graphical user interfaces (GUIs) available on commodity hardware for interactively

viewing and editing sinusoidal models. Its release was closely followed by the availability

of other GUI editors including SMSTools from Xavier Serra (running on Windows), InSpect

and ReSpect from Sylvain Marchand and Roberty Strandh (running on UNIX/X Windows),

and a number of tools from IRCAM including Diphone (Rodet and Lefèvre 1997), Sview,

and XSpect. Sadly, Lemur was never updated for newer MacOS versions. The original

version of SMSTools was discontinued and then relaunched as part of the cross-platform

CLAM project (Amatriain et al. 2002). Some of the IRCAM tools were made commercially

available as part of the IRCAM Forum software suite. In 1999, Juan Pampin developed the

ATS system which implemented sinusoidal and noise modeling in Common Lisp. A GUI

editor, ATSH by Oscar Pablo Di Liscia and Pete Moss soon followed.

1.3.2 Present Work

These packages are important predecessors of the software that will be detailed in this

document: SPEAR (Sinusoidal Partial Editing Analysis and Resynthesis). SPEAR was

created out of a desire for software that was faster, more flexible, and easier to install and

use than other options.

The following goals were kept in mind throughout the design and development phases:

editing should be as fast and as easy to understand as in a time domain waveform ed-

itor; listening to transformations should be possible with no intermediate synthesis or

processing stage; and high quality analyses should require only a few parameter set-

tings. Additionally, it was desired to have interoperability with other analysis-resynthesis

software by providing both SDIF (Wright et al. 1999a) and native format data exchange.

In order to offer a familiar and comfortable interface, SPEAR was written to run using

the native graphics of the host operating system. Development was founded on the

principle that the interface should be as close as possible to that of a first-class native

Page 22: Klingbeil Dissertation Web

1. Introduction 12

application, with little or no evidence of compatibility or GUI emulation layers. Portability

is made possible using wxWidgets (http://www.wxwidgets.org), a C++ library that allows

the software to be compiled for MacOS, Windows, and GTK Linux (Zeitlin 2001). SPEAR

was first demonstrated at the Columbia University Computer Music Center on February

23, 2004. As of this writing, builds have been created MacOS 9, MacOS X, and Windows.

The following chapters will detail the design and evolution of this software. Chapter

2 will discuss the basic techniques of spectral and sinusoidal modeling, while chapter 3

will focus specifically on sinusoidal modeling and its implementation in SPEAR. Chapter

4 details the user interface design and implementation, and chapter 5 explores some

composition applications of sinusoidal modeling.

Page 23: Klingbeil Dissertation Web

13

2. Spectral Modeling

2.1 The Frequency Domain

Spectral modeling is concerned primarily with the frequency domain content of audio

signals, as opposed to the time domain content. In the time domain, an audio signal may

be viewed as a function of time, x(t), with instantaneous amplitude as the function’s

range. Typically such a function is graphed with time on the abscissa (horizontal axis)

and amplitude on the ordinate (vertical axis), which results in the familiar waveform

representation. In the frequency domain, the same signal is represented as a new function,

X( f ), where f corresponds to frequency. Given any frequency value fk, the value X( fk)

can be used to determine relative strength of that frequency in the given signal.1 The

function X( f ) defines the spectrum of the sound.

2.1.1 Sinusoids

In order to study frequency domain representations in more detail, the notion of frequency

must be made more explicit. Any function x(t) that has a precisely repeating shape is said

to be periodic. In musical applications, and signal processing applications in general, the

periodic functions of greatest interest are the trigonometric functions sine and cosine. A

sine wave signal with a frequency of f hertz is given by the following: x(t) = sin(2π f t)

(where t is in seconds).

Because of their relation to the unit circle, the sine and cosine functions have a period-

icity of 2π. Furthermore, the cosine function may be viewed as a phase shifted form of the

1As we will define it, X( fk) is a complex value, so it does not directly give the amplitude of componentwith frequency fk. However it is straightforward to derive it from the complex value.

Page 24: Klingbeil Dissertation Web

2. Spectral Modeling 14

sine function. We use the term sinusoid to denote any any signal that can be described by a

function x(t) of the following form:

x(t) = A sin(2π f t + φ) (2.1)

where A = amplitude, f = frequency (in hertz), t = time (in seconds), and φ = initial phase

(in radians). Note that this function is periodic with a period of1f

seconds.

2.1.2 The Fourier Transform

The Fourier transform provides an explicit method for converting from the time domain

to the frequency domain. The following defines the Fourier transform of x(t):

X( f ) =∞∫−∞

x(t)e−i2π f tdt (2.2)

A complete examination of the Fourier transform is beyond the scope of this document,

but we shall consider the transform with an eye toward an intuitive and pragmatic

understanding.

Recall that x(t) represents a given time domain signal, and that evaluating the integral

for any desired frequency value f allows us to find the contribution of frequency value f in

x(t). In order to understand how frequency is defined, we must explore more closely the

inner portion of the integral, e−i2π f t. By definition, e is the base of the natural logarithm

(= 2.7182818 . . .) and i is the imaginary number√−1. Euler’s relation is a remarkable

mathematical identity that relates these constants to the sine and cosine functions:

eiθ = cos θ + i sin θ (2.3)

Note that cos θ + i sin θ is a complex value (it has both real and imaginary parts). The

expression e−i2π f t is a complex exponential that defines a complex sinusoid with frequency f .

More generally, a complex sinusoid is given by the following (compare with the definition

of a real sinusoid in equation 2.1):

x(t) = Aei(2π f t+φ) (2.4)

Page 25: Klingbeil Dissertation Web

2. Spectral Modeling 15

Complex sinusoids are particularly useful because they are able to encode phase shifts in

a manner that is compact and easy to manipulate (Puckette 2007, 189). For example, given

the complex sinusoid of equation 2.4, the properties of exponents allow us to encode both

the amplitude and the initial phase shift φ as a single complex amplitude (also known as a

phasor). Any complex amplitude A′ can take the following form (for some constant phase

value φ):

A′ = Aeiφ (2.5)

and our complex sinusoid now simplifies as follows:

x(t) = Aei(2π f t+φ)

= Aei2π f teiφ

= Aei(2π f t+φ)

= A′ei2π f t

Other important results of Euler’s formula are the following relations:

sin θ =eiθ − e−iθ

2icos θ =

eiθ + e−iθ

2

This tells us that real valued sinusoids are in fact a sum of two complex sinusoids, one

with positive frequency and the other with negative frequency. Complex sinusoids are

therefore simpler to manipulate because they have only a single frequency (either positive

or negative).

Given Euler’s identity, we may now re-express the Fourier transform as:

X( f ) =∞∫−∞

x(t)(cos 2π f t− i sin 2π f t)dt (2.6)

To gain some intuitive sense of what this implies, consider the evaluation of X( f ) for some

value fk. Informally, we may consider cos 2π fkt and sin 2π fkt to be “reference” or “basis”

sinusoids (each with freqeuncy fk) that are multiplied by the time domain signal. The

multiplication of x(t) with one of the reference function results in a new function of time,

which can be plotted as a curve. We can think of the integral as giving the total area under

Page 26: Klingbeil Dissertation Web

2. Spectral Modeling 16

the curve. The area indicates the amount of the reference sinusoid present in the signal.

X( fk) is in fact a complex amplitude that encodes amplitude and initial phase (as shown

in equation 2.5).

2.1.3 The Discrete Fourier Transform

In the given definition of the Fourier transform, the limits of integration are ±∞, implying

signals of infinite duration. In real-world situations, we work with audio signals that, by

necessity, are of finite duration and are discretely sampled. Rather than working with the

continuous Fourier transform, we can calculate a discretely sampled frequency spectrum

using the discrete Fourier transform (DFT). Given a sampled signal x(n) with a duration

of N samples, the DFT of x(n) is:

X(k) =N−1

∑n=0

x(n)e−i2π kN n k = 0, 1, 2, . . . , N − 1 (2.7)

=N−1

∑n=0

x(n)[

cos(

2πkN

n)− i sin

(2π

kN

n)]

(2.8)

Note that frequency is defined in terms of k, which takes on values from 0 to N − 1. We

can think of k as measuring the number of sinusoidal cycles per N samples. As a concrete

example, consider a signal that is sampled 44100 times per second. Let N (the length of

our signal) be 1024. If k = 1, then we have 1 cycle per 1024 samples. Converting to cycles

per second we have:

f1 =1 cycle

1024 samples× 44100 cycles

1 second≈ 43.066 Hertz (2.9)

Successive values of k correspond to 0 Hz, 43.07 Hz, 86.13 Hz, 129.20 Hz, etc. Thus, the

frequency spectrum is sampled every 43.07 Hz. Each value of k is said to refer to a specific

frequency bin. Note that the DFT gives us signal X(k) that, like the time domain signal

x(n), is N samples long.

One very useful property of the Fourier transform and the DFT is that the transform

is invertible. If we have the DFT X(k), we may recover the time domain signal x(n) as

Page 27: Klingbeil Dissertation Web

2. Spectral Modeling 17

follows:

x(n) =1N

N−1

∑k=0

X(k)ei2π kN n n = 0, 1, 2, . . . , N − 1 (2.10)

Note that the inverse DFT is very similar to the DFT, differing only in a change of sign

and a scaling factor of1N

.

As noted above, X(k) is complex valued. In general, the time domain and frequency

domain signals may be either real or complex valued. In audio applications the time

domain signal is typically real valued only, which yields complex values in the frequency

domain. Each complex amplitude X(k) = Ak encodes both the amplitude and initial

phase shift φk. From our definition of complex amplitude in equation 2.5 and from Euler’s

identity (equation 2.3), we have the following:

X(k) = Akeiφk

= Ak cos φk + iAk sin φk

Since Ak and φk are constants, we can represent our complex amplitude as a complex

number of the form a + bi where a = Ak cos φk and b = Ak sin φk. We can visualize this as

a point (a, b) on the complex plane. The amplitude Ak of the sinusoid is the distance from

the origin to the point (a, b), and the phase φk is the angle of rotation about the origin. The

amplitude Ak is defined as the magnitude of the complex point.

Ak = |X(k)| =√

a2 + b2 (2.11)

The phase angle φk is given by the inverse tangent function.

φk = tan−1 ba

(2.12)

Equations 2.11 and 2.12 give an explicit method to convert from a complex frequency bin

value to magnitude (amplitude) and phase. Note that for a real valued signal, phase is

measured relative to a cosine function. Phases may be expressed in sine phase via a phase

shift of −π2 radians.

Page 28: Klingbeil Dissertation Web

2. Spectral Modeling 18

2.1.4 The Short-Time Fourier Transform

While almost all musical signals have spectra that vary significantly over time, many

sounds exhibit spectra that are locally stable. Such signals may be well-represented by a

series of spectral “snapshots” or “frames” that capture the average spectral content and, in

sequence, represent the time varying spectrum of the sound. Such an analysis is called a

short-time Fourier transform (STFT). Each spectral frame is computed by means of a DFT

of length N and is centered on a particular sample n. Successive frames are computed by

advancing some number of samples H (the hop size) and computing another DFT.

As noted above, the DFT is a discrete sampling of the frequency spectrum. When a

signal consists of a sinusoid with a period that is an integer factor of the DFT length N,

then the sinusoid’s frequency is precisely aligned with one of the DFT frequency bins.

Figure 2.1 shows the magnitude spectrum of a DFT (length N = 16) of a sinusoid with

period N4 . Note that the magnitude spectrum consists of a single spike at bin 4.

Figure 2.1: DFT magnitude spectrum of a sinusoid with period N4 , (N = 32).

Since such precisely constructed signals are unlikely to be found in real-world digital

recordings, let us now consider the case of a sinusoid with a period of N4.21 samples, and

its corresponding DFT magnitude spectrum (figure 2.2). In this case, there are significant

components present in bins 0 through 16. Since the signal is N samples long, but has a

sinusoidal period of 4.21 cycles, it cuts off abruptly creating a significant discontinuity.

Page 29: Klingbeil Dissertation Web

2. Spectral Modeling 19

Figure 2.2: DFT magnitude spectrum of a sinusoid with period N4.21 , (N = 32).

Figure 2.3 compares two sinusoids, one with period N4 and the other with period N

4.21 . Note

that the sinusoid with period of N4 arrives smoothly at 0 at sample index 32.

Figure 2.3: Sinusoids of 4.21 cycles and 4.0 cycles per N samples, (N = 32).

Intuitively, one can interpret such discontinuities as additional high frequency com-

ponents. These additional components are often described as spectral distortion or spectral

leakage. These distortions are a significant issue when one is attempting to detect multiple

sinusoids at different frequencies: it may be impossible to differentiate spectral leakage

from sinusoids. By applying a smoothing window function, h(m), in the time domain, these

edge discontinuities can be reduced and the spectral leakage attenuated. The STFT of

Page 30: Klingbeil Dissertation Web

2. Spectral Modeling 20

length N beginning at sample n and with smoothing window h(m) is defined as follows:

Xn(k) =N−1

∑m=0

x(m + n)h(m)e−i2π kN m (2.13)

A number of different window functions are commonly used in audio applications.

The most effective are generally bell-shaped curves. Note that the apparent absence of a

smooth window function in fact implies a rectangular window function. The rectangular

window function is defined to be 1 on the interval 0 . . . N − 1 and 0 otherwise. See Harris

(1978) and Nuttall (1981) for a complete discussion of window functions, their definitions,

and properties.

Before continuing further with our discussion of windowing, we must introduce one

other important signal processing concept. The convolution operation ∗ for two signals

x(n) and y(n) is defined as follows:

f (n) = x(n) ∗ y(n) =N−1

∑m=0

x(m)y(n−m) (2.14)

The convolution of x(n) and y(n) can be thought of as a new signal, f (n), where each

sample of x is multiplied by a time shifted copy of all the samples in y and the results

summed. Thus, f (n) consists of the scaled shifted sums of copies of y(n).

The convolution theorem (which we will not prove here) states that the spectrum of the

convolution of two signals is equivalent to the multiplication of the signals each expressed

in the frequency domain:

DFT [x(n) ∗ y(n)] = DFT [x(n)] DFT [y(n)]

= X(n)Y(n)

The converse is also true: the spectrum of the multiplication of two signals in the time

domain is equivalent to the convolution of the signals each expressed in the frequency

domain:

DFT [x(n)y(n)] = DFT [x(n)] ∗DFT [y(n)]

= X(n) ∗Y(n)

Page 31: Klingbeil Dissertation Web

2. Spectral Modeling 21

By the convolution theorem, the multiplication of a signal x(n) and a window h(n) is

equivalent to their convolution in the frequency domain: DFT [x(n)h(n)] = X(n) ∗ H(n).

Since the spectrum of a sinusoid is a single impulse, it follows that the spectrum of a

windowed sinusoid will be a scaled copy of the spectrum of the window function, centered

on the frequency of the sinusoid. Returning to the example of a sinusoid with a period of

4.21 cycles per window, we can get a sense of the effects of windowing by inspecting the

underlying continuous spectrum.2

Figure 2.4: DFT magnitude spectrum and the interpolated underlying magnitude spectrumof a sinusoid with period N

4.21 , (N = 32).

The dotted curve in figure 2.4 shows the continuous underlying spectrum which

consists of a main lobe centered at the sinusoid frequency and many side lobes of lower

amplitude arranged around the main lobe. Different window functions exhibit unique

main lobe shapes and side lobe levels. The case of a sinusoid with an integer period

and a rectangular window is quite special. Figure 2.5 shows that the although there are

significant side lobes in the underlying spectrum, the DFT samples align precisely with

nulls in the continuous spectrum. The result is a single impulse in frequecy bin 4.

Examining the spectra of various window functions can provide a good sense of their

practical capabilities. Figures 2.6 through 2.9 show various window functions in both

the time domain and frequency domains. Note that as the center area of the window

2The continuous spectrum can be approximated by zero-padding the signal and then performing the DFT.The result is a spectrum that is oversampled by the zero-padding factor.

Page 32: Klingbeil Dissertation Web

2. Spectral Modeling 22

Figure 2.5: DFT magnitude spectrum and the interpolated underlying magnitude spectrumof a sinusoid with period N

4 , (N = 32).

function becomes narrower in the time domain, the main lobe of the window spectrum

becomes wider. This demonstrates an important tradeoff in the design and use of window

functions.

By using the STFT and appropriate windowing, we can successfully model many types

of musical signals and more significantly, we can use these models to effect a number of

different sonic transformations.

2.2 STFT Analysis and the Phase Vocoder

The phase vocoder is one of the most widely utilized STFT techniques for time and

frequency scale modification. The technique makes use of phase information in successive

analysis frames to make a precise estimate of the sinusoidal frequency present at a

particular analysis bin. By default, the frequency resolution of the STFT is limited by

the bin spacing, which in turn is controlled by the DFT length N. For example, as was

observed in equation 2.9, a DFT of length 1024 at a sampling rate of 44.1 kHz results in

a bin resolution of 43.07 Hz. Practical DFTs are computed by means of the fast Fourier

transform which limits N to powers of 2. Therefore, there are some significant restrictions

on the frequency resolution. Moreover, the center frequencies of each STFT bin will be

limited to harmonics of the center frequency of bin 1.

Page 33: Klingbeil Dissertation Web

2. Spectral Modeling 23

Figure 2.6: Time domain plot and magnitude spectrum of the rectangular window.

Figure 2.7: Time domain plot and magnitude spectrum of the triangular window (N = 31).

Figure 2.8: Time domain plot and magnitude spectrum of the Hann window.

Figure 2.9: Time domain plot and magnitude spectrum of the Blackman window.

Page 34: Klingbeil Dissertation Web

2. Spectral Modeling 24

Purely harmonic sounds may be analyzed by resampling the signal in the time domain

so that the length of the fundamental period is an integer factor of the DFT length N

(Beauchamp 1993). This assures that each harmonic in the input will be aligned with the

center of an analysis bin. However, this does not accommodate inharmonic sounds, nor

does it allow the measurement of frequency deviations (including vibrato) that are present

in almost all real-world acoustic signals.

2.2.1 Putting Phase to Good Use

Thus far we have been looking at the magnitude spectrum, but each analysis frame also

has a phase spectrum and this information is useful. Since frequency is defined as the

amount of phase change per time unit, the change in phase between successive analysis

frames can be used directly to estimate frequency. We define the phase deviation for

frequency bin k and analysis frame n as follows:

∆φk,n = φk,n −[

φk,n−1 + H2πk∆ f

fs

](2.15)

where φk,n is the phase for bin k, frame n, H is the analysis hop size in samples, ∆ f is the

frequency bin spacing in hertz, and fs is the sampling rate. Equation 2.15 expresses the

idea of phase unwrapping, where the term φk,n−1 + H 2πk∆ ffs

represents the unwrapped

phase of a sinusoid with frequency centered precisely on analysis bin k. Positive phase

deviation ∆φk,n indicates a sinusoid with a frequency higher than the bin center frequency,

negative phase deviation indicates lower frequency.

In practice, the phase deviations must be re-wrapped to the half-open interval [−π, π)

which constrains the allowable frequency deviation and locates the estimated frequency

around analysis bin center frequency. If Θφk,n represents the re-wrapped phase deviation,

then the estimated frequency (in hertz) for bin k and analysis frame n is

fk,n =[

Θφk,n

] [fs

H

]+ k∆ f (2.16)

Page 35: Klingbeil Dissertation Web

2. Spectral Modeling 25

Note that decreasing the hop size increases the range of possible frequency deviation. Small

hop sizes will allow the phase vocoder to more closely track the rapid phase deviations of

non-stationary (noisy) spectral components.

For quasi-stationary (sinusoidal) components, we note that there is an important

relationship between the choice of window function and hop size. As mentioned earlier, a

single sinusoidal component will result in a spectrum that is a scaled translation of the

spectrum of the analysis window. As such, we desire that all bins in the main lobe of

the window should contribute to the energy of the sinusoid, i.e. their frequencies should

“lock” to the frequency at the center of the main lobe. For a main lobe width of W bins and

window size of M, a hop size H such that H ≤ MW will be sufficient (Puckette 1995). Thus,

the wider the main lobe, the smaller the allowable hop size. This also makes intuitive sense

from a time domain perspective. Since wider main lobes result from narrower bell-like

window shapes in the time domain, the decreased width of the window shape must be

compensated for by more dense sampling in the time domain.

2.2.2 Resynthesis and Transformation

Following analysis, the phase vocoder data can be used to perform a resynthesis of the

original sound. Depending on the desired result, the analysis data may be modified to

effect different sonic transformations. Two different resynthesis methods are in common

use.

The first is the STFT overlap-add technique. With this method, the inverse DFT is

applied to the data in each analysis frame. Since the analysis hop size Ha is typically less

than the DFT length N, a time domain overlapping of each DFT window results. Provided

that the analysis window function sums to unity upon overlap (which can be assured

by choosing the appropriate window function and normalization factor), the original

signal can be recovered exactly. Setting the hop size so that Hs 6= Ha gives a time warped

resynthesis.

Page 36: Klingbeil Dissertation Web

2. Spectral Modeling 26

The second resynthesis method is the summing oscillator bank. Each bin of the

phase vocoder data is associated with a sinusoidal oscillator and each phase vocoder

frame provides frequency, amplitude, and phase information that can be used to drive

the oscillators. Since each analysis frame is separated by a hop of Ha samples, linear

interpolation of frequency and amplitude is often used to fill in intermediate samples. With

linearly interpolated frequency, the phase function becomes the integral of instantaneous

interpolated frequency. As such, the resynthesized phases cannot be guaranteed to match

the original phases, and thus resynthesis with the oscillator bank will not recover the

original signal. In practice, different frequency interpolation strategies can be used to

achieve fairly high fidelity results. For example, the cubic phase interpolation method

commonly used in sinusoidal modeling can be used to match phase at each frame (see

section 3.3.4 for details).

Typically one is interested in transforming the analysis data in order to achieve one or

more of the following results:

1. frequency filtering (boosting or cutting certain frequency bands)

2. time expansion or compression

3. transposition (frequency scaling)

Of these three, frequency filtering is the most straightforward, and can easily be imple-

mented by selectively scaling the amplitudes in the desired analysis bins. Time varying

filtering may be used in applications such as broadband noise reduction.

2.2.3 Time Expansion/Compression

Time expansion/compression may be achieved in the overlap add case by adjusting Hs

and/or Ha. The expansion ratio is given by α =Hs

Ha. When performing time expansion or

compression, care must be taken to properly update the phases in each synthesis frame.

Let φ′k(n) denote the resynthesis phase for bin k in frame n. Phase is propagated forward

Page 37: Klingbeil Dissertation Web

2. Spectral Modeling 27

from each previous frame according to the analysis frequencies as follows:

φ′k(0) = αφk(0) (2.17)

φ′k(n) = φ′k(n− 1) + Hsfk(n)2π

fs(2.18)

Note that the initial phase is multiplied by the expansion ratio α. In their 1997 study,

Laroche and Dolson show that scaling the initial phases will preserve the original phase

relationships when expanding by an integer factor.

Because time expansion/compression no longer guarantees the same phase relation-

ships as the original sound, an additional smoothing synthesis window function is gen-

erally applied after each inverse DFT. This serves to reduce the effect of any phase or

amplitude discontinuities that result from frequency domain transformations. Since the

application of a synthesis window amplitude modulates the signal, proper care must be

taken to assure enough overlap to reduce amplitude modulation artifacts.

To explore this issue in more detail, it will prove useful to examine the amplitude

behavior of several analysis and synthesis windows under different overlap-add conditions.

First let us define the overlap factor β =NHs

. When β = 1 there is no overlap, and the

output consists of a signal that is amplitude modulated by the product of the synthesis

window with the analysis window. (The original analysis window was recovered from

the inverse DFT although the shape may be distorted due to operations performed in

the frequency domain.) Figure 2.10 shows the overlap-adding of Blackman analysis and

synthesis windows with β values of 1, 3, 4, and 6. Note that un-modulated gain is achieved

when β = 6. In general, windows with a narrower time domain peak will require more

overlap.

Additional steps must be taken to assure proper amplitude normalization. For a

synthesis window of w′ of length N, the normalization factor γ is

γ = Hs

N−1

∑i=0

w′(i) (2.19)

Page 38: Klingbeil Dissertation Web

2. Spectral Modeling 28

Figure 2.10: Overlap-adding the product of Blackman analysis and synthesis windowswith varying overlap factors (β).

Time varying expansion and contraction factors imply a variable synthesis hop size

with a corresponding variable normalization factor. Due to the possibility of amplitude

modulation artifacts, care must be taken not to increase Hs beyond the point of constant

gain. In practice it is best to fix Hs and vary Ha (or linearly interpolate frequency and

amplitude from the analysis spectrum) to achieve the desired expansion ratio.

2.2.4 Transposition

Transposition (frequency scaling) with the phase vocoder may be handled in a number of

different ways. A simple, though computationally expensive, approach synthesizes each

phase vocoder band with a sinusoidal oscillator. Frequency and amplitude may be linearly

interpolated from frame to frame. If Fk(n) and Ak(n) are the piecewise interpolated

frequency and amplitude functions for bin k, then the resynthesis is

x(n) =N/2−1

∑k=0

Ak(n) cos(

nPFk(n)2π

fs+ φk

)(2.20)

where P is the transposition ratio, and φk is the initial phase of bin k.

Page 39: Klingbeil Dissertation Web

2. Spectral Modeling 29

A less computationally expensive approach is to resynthesize with the overlap-add

method with a time expansion/compression ratio that is the inverse of the transposition

ratio. The transposition is then achieved via resampling in the time domain. Another

approach is to scale the spectrum in the frequency domain as described in Laroche and

Dolson (1999). The spectral magnitude peaks are scaled according to the transposition

ratio, and the phases are adjusted to reflect the new transposed frequencies.

2.2.5 Refinements

A number of additional refinements can be added to the basic phase vocoder technique.

One drawback that has been observed with time expansion/compression is the problem of

“phasiness.” This can best be described as a loss of presence in the synthesized signal; the

result is less crisp and there is sense of greater reverberation. Much of this can be attributed

to a lack of phase coherence between adjacent frequency bins. Although the standard

phase vocoder technique interpolates coherent phase between adjacent analysis frames

(“horizontal” phase), it does nothing to maintain “vertical” phase alignment. Puckette

(1995) proposed a method of phase locking that proceeds by first finding amplitude peaks

in each frame. Under the assumption that frequency bins adjacent to the peak contribute

to the same sinusoid, the phases of adjacent bins are set equal to the phase of the peak bin.

Laroche and Dolson (1997) describe a number of different phase maintenance strategies (all

based on this peak detection method) that can provide significant improvements in sound

quality. Phase relationships are particularly important for preserving transient and micro-

transient events. Robel (2003) describes a technique for reinitializing phases at transients

that significantly reduces phasiness and transient smearing under time scale modifications.

Although the phase vocoder has its limitations, when carefully implemented it can offer

excellent fidelity.

Page 40: Klingbeil Dissertation Web

2. Spectral Modeling 30

2.3 Spectral Envelopes

Up to this point we have been discussing the precise frequency domain content of audio

signals. It has been shown that the ear functions similarly to Fourier analysis, with specific

locations along the basilar membrane of the inner-ear corresponding to specific frequencies.

However, in addition to frequency content, we are also simultaneously sensitive to the

overall general shape of the spectrum. We call the curve that defines the shape of the

magnitude spectrum the spectral envelope. The dotted curve in figure 2.11 shows one

possible spectral envelope for the given spectrum.

Figure 2.11: Spectral envelope of a soprano voice.

Many vocal and instrumental sounds can be understood in terms of a source-filter model,

where the resonant body of the instrument or of the vocal tract shape (filter) the input (the

source). In voiced human speech, the source is provided by the periodic vibration of the

vocal folds. The spectral envelope can be taken as a model of the resonances of the vocal

tract. For vowels, the peaks in the spectral envelope trace the formants. For any particular

speaker and vowel, the spectral envelope stays fixed regardless of pitch.

If a sound is transposed using the phase vocoder, for example, the spectral envelope

shape will change along with the transposed spectrum. For an instrumental or vocal

Page 41: Klingbeil Dissertation Web

2. Spectral Modeling 31

sound, the perceived result is a change in size of the resonating body. Thus there is

a change in timbre as well as in pitch. The timbre can be preserved by dividing out

the spectral envelope prior to transposition (this process of flattening the spectrum is

sometimes referred to as whitening), and then filtering the transposed magnitude spectrum

by the original spectral envelope.

Common methods for determining the spectral envelope include linear predictive

coding (LPC), cepstrum, discrete cepstrum, and true envelope methods (Robel and Rodet

2005). For a thorough discussion of spectral envelope estimation methods and applications

see Schwarz (1998).

2.4 Sinusoidal Modeling

Like the phase vocoder, sinusoidal modeling is an STFT based analysis/resynthesis

technique. It attempts to represent an audio signal as a generalized collection of sinusoids

with time varying frequency, amplitude, and phase. Unlike the phase vocoder, the number

of sinusoids present at a given time is not fixed by the DFT size. As a consequence, the

sinusoids need not necessarily be centered around a fixed frequency band, and sinusoids

may vary arbitrarily in frequency. The following plots of a frequency sweep (chirp) first

analyzed with the phase vocoder (figure 2.12) and then with sinusoidal modeling (figure

2.13), clearly illustrate the difference between the two approaches.

The phase vocoder represents the chirp with many frequency bands that each lock

to the chirp frequency as it sweeps upward. The result is highly overspecified, with

many bands per sinusoidal feature. Manipulation of the chirp (transposition for example)

requires the adjustment of many frequency bands. As a corollary, a local adjustment to

any one frequency band will degrade the integrity of the frequency sweep.

In contrast, a sinusoidal model represents the chirp as a single breakpoint function

with time varying frequency. The chirp can be freely transposed and reshaped by the

manipulation of a single frequency/amplitude trajectory. Although in most cases perfect

Page 42: Klingbeil Dissertation Web

2. Spectral Modeling 32

Figure 2.12: Phase vocoder analysis of a frequency sweep from 100 Hz to 500 Hz.

Figure 2.13: Sinusoidal modeling analysis of a frequency sweep from 100 Hz to 500 Hz.

reconstruction is not possible with a purely sinusoidal model, the benefits of the technique

may outweigh this drawback.

2.4.1 STFT Frequency Interpolation

The basic method of sinusoidal modeling begins with a STFT. As observed in section

2.1.4, the short-time spectrum of a signal is the convolution of the spectrum of the window

function with the spectrum of signal x(n). If the signal in question is in fact a single

sinusoid at frequency fk, then its magnitude spectrum is a single impulse peak at frequency

fk, and its windowed magnitude spectrum is simply the spectrum of the window function

Page 43: Klingbeil Dissertation Web

2. Spectral Modeling 33

with its main-lobe centered at frequency fk. In the DFT magnitude spectrum, a sinusoid

appears as a local maximum near the true maximum of the underlying continuous

spectrum. Parabolic interpolation may then be used to approximate the position of the

peak of the main lobe.

αβ γ

estimated peak

Figure 2.14: Estimating a peak in the spectrum based on a parabolic curve fitting.

Given the DFT X(k), the magnitude spectrum can be searched for bins that are local

maxima. Bin n is considered a local maximum if its magnitude is greater than its two

neighboring bins, or more precisely |X(kn−1)| < |X(kn)| > |X(kn+1)|. We can then fit

a parabola that passes through the magnitude values of bins n− 1, n, and n + 1. The

parabola takes the following form:

y(x) = a(x− p)2 + b

The known points on the curve are y(−1), y(0), and y(1) and are given by the the bin

magnitudes (measured in dB). According to Smith and Serra (1987), frequency estimates

are found to be more accurate when interpolating using dB rather than linear magnitude.

y(−1) = α = 20 log10 |X(kn−1)|

y(0) = β = 20 log10 |X(kn)|

y(1) = γ = 20 log10 |X(kn+1)|

Page 44: Klingbeil Dissertation Web

2. Spectral Modeling 34

With three equations and three unknowns (a, p, and b) we can solve for the parabola peak

location p.

p =12

(α− γ

α− 2β + γ

)(2.21)

The center frequency (measured in bins) will be n + p. The parabola peak location p can

then be used to solve y(p) for both the real and imaginary parts of the complex spectrum.

For the real part, we set α<, β<, and γ< to the real values at bins n− 1, n, and n + 1,

α< = <e[X(kn−1)]

β< = <e[X(kn)]

γ< = <e[X(kn+1)]

and for the imaginary part

α= = =m[X(kn−1)]

β= = =m[X(kn)]

γ= = =m[X(kn+1)]

Solving y(p) for the real part a and imaginary part b gives

a = y(p)< = β< −14

p(α< − γ<)

b = y(p)= = β= −14

p(α= − γ=)

According to equations 2.11 and 2.12 the magnitude wil be√

a2 + b2 and the phase will be

tan−1 ba

.

This frequency, phase, and magnitude interpolation method means that the analysis

procedure can accurately detect sinusoidal frequencies that are non-integer multiples

of the analysis bin spacing frequency. This makes it ideal for analyzing sounds that

have inharmonic components, have variable frequency content, or exhibit limited non-

stationary behavior (a sum of different harmonic complex tones, sounds with wide vibrato

or glissandi, etc.)

Page 45: Klingbeil Dissertation Web

2. Spectral Modeling 35

2.4.2 Partial Tracking

Once all of the sinusoidal peaks in a particular STFT frame have been detected, the

peaks must be organized such that they form continuous, time-varying sinusoidal tracks

(partials). This is accomplished by matching the peaks in the current frame with peaks

present in previous STFT frames. During resynthesis the frequency, amplitude, and phase

can be smoothly interpolated from a peak to its matched successor in the next frame.

k0 k1 k2 k3 k4 k5

time

frequ

ency

unmatched peak

track start (birth)

track end (death)

frequency threshold

Figure 2.15: Connecting STFT peaks into sinusoidal tracks.

Typically, some peaks that fall below a certain amplitude threshold will be discarded

prior to matching. This means that each frame may contain different numbers of peaks and

that, over time, individual partials may vanish or appear depending on their amplitude.

Amplitude thresholding techniques and tradeoffs will be discussed in more detail in the

next chapter. The major implication for this type of sinusoidal model is that the number

of sinusoids present at any one instant is variable. By eliminating redundancies and low

amplitude components, the result is a model that is efficient to store and manipulate.

Several different techniques have been proposed for partial tracking. The simplest

method is a locally optimal greedy algorithm that matches each peak to the peak in

the next frame that is closest in frequency. Typically there is a distance threshold that

constrains the maximum allowable frequency jump. Each peak is defined in terms of the

following data structure:

Page 46: Klingbeil Dissertation Web

2. Spectral Modeling 36

structure Peak

FreqAmpPhaseForwardmatchBackmatch

end structure

Note that peaks keep track of both their successor (Forwardmatch) and predecessor

(Backmatch). Forwardmatch and Backmatch are initialized to Nil, indicating the absence

of a matching peak. Listing 2.1 illustrates the greedy peak matching algorithm.

1: for prevpeak ∈ Pk−1 do2: for curpeak ∈ Pk do3: distance← |prevpeak.Freq− curpeak.Freq|4: if distance < f req_threshold then5: if curpeak.Backmatch 6= Nil then6: existing_distance← |curpeak.Backmatch.Freq− curpeak.Freq|7: else8: existing_distance← f req_threshold9: end if

10: if distance < existing_distance then11: curpeak.Backmatch.Forwardmatch← Nil

12: curpeak.Backmatch← prevpeak13: prevpeak.Forwardmatch← curpeak14: end if15: end if16: end for17: end for

Algorithm 2.1: Greedy peak matching from frame k− 1 to frame k.

Pk is the collection of peaks in frame k (the current frame) and Pk−1 is the collection of

peaks in the previous frame. The outer loop iterates through each of the peaks in frame

k− 1. The inner loop looks for the best match in the current frame by iterating through

each of frame k’s peaks. If the frequency distance between any two peaks is less than

distance_threshold (line 4), then it is a candidate for matching. Before connecting the peaks,

the algorithm checks to see if the candidate peak in frame k + 1 has already been claimed

(line 5). If so, the new distance must be less than distance of the current match (line 10).

Page 47: Klingbeil Dissertation Web

2. Spectral Modeling 37

Following matching, the Forwardmatch and Backmatch fields indicate the complete

sinusoidal tracks. If Backmatch = Nil, a new partial is starting up. If Forwardmatch = Nil,

a partial is ending. When a partial begins or ends, a short amplitude fade is usually added

in order to avoid a discontinuity.

Given that the greedy algorithm has no explicit model of the underlying sound

structure, nor any context beyond the peaks of the current and previous frames, it actually

performs reasonably well in practice. However, the greedy algorithm can be easily confused

by spurious peaks or significant frequency deviations such as vibrato or glissandi (see

figure 2.16). A number of minor refinements can be made, such as incorporating amplitude

into the measure of peak proximity or measuring frequency distance on a logarithmic

scale.

k0 k1 k2 k3 k4 k5

time

frequ

ency

partial tracking conflicts

Figure 2.16: Partial tracking conflicts due to glissandi.

2.4.3 Improved Partial Tracking

A number of other improved partial tracking strategies have been proposed. Under the

assumption that the sound being analyzed contains relatively stable frequency trajectories

(such as stable harmonic or inharmonic partials of a single fundamental), peaks can be

matched to a set of frequency guides (Serra 1989). Guide based partial tracking proceeds

as follows: at frame k we have a set of N peaks Pk with frequencies f0, f1, f2 . . . fN−1,

and a set of M frequency guides G with frequencies f ′0, f ′1, f ′2 . . . f ′M−1. As in the greedy

Page 48: Klingbeil Dissertation Web

2. Spectral Modeling 38

algorithm, each guide attempts to claim the peak in Pk that is closest in frequency to the

guide frequency (| f ′m − fn| is minimized for all combinations, subject to constraints on

the maximum allowable frequency jump). When a match is found between partial n and

guide m, the partial trajectory is continued and a new guide frequency fm is computed

according to

fm = α( fn − f ′m) + f ′m for α ∈ [0, 1]

The parameter α is a simple lowpass filter on the guide frequency. When α is zero, the

guide frequencies remain constant. As α approaches 1, the guide frequencies more closely

track the frequency of the evolving partial. Guides that fail to find a match are set to a

dormant state. They can “wake up” in future frames if they find a peak appropriately

proximate in frequency. If a guide remains dormant for too long, it is removed from G.

Peaks that are not matched to a guide may (optionally) spawn new guides. The number of

maximum allowable guides may be limited, if desired.

For harmonic sound, the guide frequencies may be fixed to multiples of the fundamen-

tal frequency f0. In this case the number of guides can remain constant throughout the

evolution of the sound. Note that the tracking method works well only if the fundamental

frequency estimation is reliable. Guide based tracking has two distinct advantages: it can

fill temporal gaps in the frequency trajectory, and it can provide a clean representation

for harmonic sounds (with one partial per harmonic). Guide based tracking has been

implemented in a number of sinusoidal modeling packages including SMS, Loris, and

ATS.

For sounds with non-stationary behavior (noise, microtransients, rapid frequency

fluctuations, etc.) guide based tracking offers few advantages in comparison to the basic

greedy algorithm. Depalle, García, and Rodet proposed a tracking method based on

a combinatorial Hidden Markov Model (HMM) that provides more globally optimal

frequency trajectories (Depalle et al. 1993). A Hidden Markov Model consists of a set

of related states, observations, and parameters. In a typical HMM problem, the goal is

to determine a temporal sequence of states based solely on observations that have some

Page 49: Klingbeil Dissertation Web

2. Spectral Modeling 39

probabilistic connection to the states. The states are the “hidden” part of the model,

and the parameters determine the probability of a state (or state transition) given certain

observations.

In the application of HMMs to partial tracking, a state consists of a set of connections

from frame k− 1 to k. The goal is to find the most likely sequence of states (connections)

given the observations. In this case the observation is simply the number of peaks present

in frame k− 1 and in frame k. If we let Sk model the set of peak connections from frame

k− 1 to k then next state, Sk+1 will specify the connections from frame k to k + 1. Thus,

a state transition involves the peak trajectories of three frames: k− 1, k, and k + 1. The

parameters of the HMM specify the probability of a transition from state to state. The

parameters are based on a cost function that favors peak connections with continuous

frequency slopes. Note that actual frequency distance is not considered. For this reason

the HMM method can be used to track crossing partials. Given the parameters of the

model (the probability of a particular state transition), the Viterbi algorithm is used to find

the most likely sequence of states. The optimal sequence of transitions will be that which

minimizes frequency discontinuities.

The computational cost of the HMM method is high since it must consider all com-

binations of peak connections across all frames. In practice, the number of frames (and

therefore the number of state transitions) is limited to some window of T frames. The

number of possible peak connections under consideration is further constrained by fixing

the number of partials alive during any interval of T. Other options could further constrain

the combinations, such as limiting maximum allowable frequency slopes or limiting the

number of partial crossings.

A particular weakness of the HMM method described by Depalle et al. is that it cannot

fill temporal gaps in the partial trajectories. Gaps may occur (especially in low amplitude

partials) due to bin contamination from noise, or if one is analyzing material that has

undergone lossy compression (a situation that may be common given current audio

distribution methods). Two new promising partial tracking approaches include future

trajectory exploration (Lagrange et al. 2004) and tracking using linear prediction (Lagrange

Page 50: Klingbeil Dissertation Web

2. Spectral Modeling 40

et al. 2003). SPEAR has implemented the linear prediction method which will be detailed

in section 3.2.

2.4.4 Synthesis

The standard resynthesis method for a sinusoidal model is the summing oscillator bank.

Each frequency trajectory computed via partial tracking is represented as a breakpoint

envelope where where point k of envelope i is represented by the time, frequency, ampli-

tude, and phase values (tki , f k

i , aki , φk

i ). Linear interpolation of amplitude and frequency

can be used to resynthesize each breakpoint segment. Fki (n) and Ak

i (n) are the linear

interpolation frequency and amplitude functions for the segment from breakpoint i to

breakpoint i + 1.

Fki (n) = f k

i +(

f ki+1 − f k

i

)( nfs− tk

i

tki+1 − tk

i

)for fstk

i ≤ n < fstki+1 (2.22)

Aki (n) = ak

i +(

aki+1 − ak

i

)( nfs− tk

i

tki+1 − tk

i

)for fstk

i ≤ n < fstki+1 (2.23)

where n is the output sample index, fs is the sampling rate, and tki is breakpoint time in

seconds. If we let Fi(n) and Ai(n) be the piecewise amplitude and frequency functions for

partial i defined for all n, then the final resynthesis is given by

x(n) =N−1

∑i=0

Ai(n) cos(

nFi(n)2π

fs+ φi

)(2.24)

where φi is the initial phase of partial i.

Note that the instantaneous phase is the integral of the frequency function. For linear

frequency interpolation this implies a quadratic phase function. The phase values at all

but the first breakpoint are ignored. If we desire a resynthesis that is phase accurate, a

different interpolation strategy must be used. McAulay and Quatieri (1986) developed a

cubic phase interpolation method that matches frequency and phase at each breakpoint.

This will be discussed in detail in section 3.3.4.

Page 51: Klingbeil Dissertation Web

2. Spectral Modeling 41

2.4.5 Limitations and Extensions

Sinusoidal modeling makes the assumption that the audio signal can be closely approxi-

mated as a sum of sinusoids. Although this may hold true for certain classes of signals

(harmonic sounds with clear tonal components and slowly varying frequency content)

many types of signals are not well represented by sinusoids — broadband noise or transient

events such a snare drum hit, for example. In many cases, sounds of interest will exhibit

both tonal and noisy components. For example, a flute tone will contain quite a bit of

broadband breath noise in addition to a clear fundamental.

Partial tracking becomes particularly difficult in the presence of noise, reverberation,

or dense polyphony. Recent research continues to suggest new approaches for partial

tracking, noise modeling, and transient preservation. A few of these methods have been

implemented in SPEAR and will be discussed in the next chapter.

Even with extensions to the basic sinusoidal modeling method, perfect signal recon-

struction is not guaranteed. However, when considering compositional goals, perfect

reconstruction is rarely the intent. If lossless reproduction is required, sampling synthesis

accomplishes this with far greater ease and efficiency. The power of a sinusoidal model

lies in its potential to allow independent transformations in both time and frequency. One

can begin with models that closely mimic real-world acoustic events and then, through

a series of transformations or interpolations, explore a vast space of synthetic variations.

Clearly this is fertile territory for creative musical exploration.

Page 52: Klingbeil Dissertation Web

42

3. Sinusoidal Analysis and Resynthesis

SPEAR provides two different spectral analysis modes — the phase vocoder and sinusoidal

modeling. The phase vocoder follows the standard method described in the preceding chap-

ter. This chapter will focus on specific implementation of sinusoidal analysis/resynthesis.

Several extensions to the basic techniques, as described in the preceding chapter, will be

discussed.

3.1 Analysis Parameters

Sinusoidal analysis requires a number of input parameters including window type, window

size, DFT size, analysis hop size, amplitude thresholds, and partial tracking constraints.

All of these parameters can have a significant effect on the quality of the analysis. The

interaction of amplitude thresholds and partial tracking can be particularly complex. The

goal is to provide defaults for all of these parameters that work reasonably well for a

wide variety of musical signals. However, even with careful choices, some user-selected

parameter inputs are almost always required. Typically, the user might begin analysis

using the default settings and then adjust accordingly after viewing the analysis and

auditioning the resynthesis.

3.1.1 Analysis Window

One of the most crucial parameter choices is the length of the analysis window. This will

control width of the main lobe in the frequency domain and, as a result, will determine the

frequency resolution of the analysis. Rather than specifying the window length in samples,

the user inputs the frequency resolution fa in hertz. Adjacent sinusoids that are at least fa

Page 53: Klingbeil Dissertation Web

3. Sinusoidal Analysis and Resynthesis 43

hertz apart can be properly resolved in the analysis. Additionally, no significant sinusoid

with a frequency less than fa hertz should be present in the input. For a harmonic sound,

fa should correspond to the frequency of the fundamental.

Given fa, we must determine the appropriate window size and DFT size. In order to

resolve sinusoids using the peak picking and quadratic interpolation method, the window

size must be large enough to insure separation of the window main lobes. The window

length M is given by

M = ∆s(

fs

fa

)(3.1)

where fs is the sampling rate and ∆s is the desired window separation in bins.

SPEAR uses a Blackman window which has a main lobe width of 6 bins. If we require

complete separation of the main lobes as shown in figure 3.1, then ∆s = 6. Note that as

∆s increases, the window length grows. Since large windows introduce an undesirable

loss of temporal resolution, we would like to use a smaller value of ∆s that still allows

for the resolution of closely spaced sinusoids. Abe and Smith (2004) made an extensive

study of the minimum allowable frequency separation (MAFS) for various window types.

The MAFS is the smallest separation that allows for peak detection and does not introduce

significant bias in the peak interpolation due to interference from a neighboring peak. With

a spectral oversampling (zero padding) factor of 2, the MAFS for the Blackman window is

3.53 bins (Abe and Smith 2004, 8, table 9). This is shown in figure 3.2. SPEAR uses the

slightly more conservative separation of 4 bins. Given a window length M, the FFT size is

N = 2dlg Me+1, which results in spectrum oversampling by a minimum factor of 2.

With a peak separation of 4 fa and spectrum oversampling by a minimum factor of

2, each main lobe peak will be sampled by at least 7 DFT bins. To aid the rejection of

spurious peaks we may conservatively impose the restriction that the three magnitudes

|X(kn−1)|, |X(kn)|, |X(kn+1)| of a parabolic peak must also all exceed the magnitude of

either neighboring bin: |X(kn−1)| > |X(kn−2)| or |X(kn+1)| > |X(kn+2)|.

Page 54: Klingbeil Dissertation Web

3. Sinusoidal Analysis and Resynthesis 44

Figure 3.1: Magnitude spectrum of the main lobes of two Blackman windows with afrequency separation ∆s = 6.

Figure 3.2: Magnitude spectrum of the main lobes of two Blackman windows with theminimum allowable frequency separation ∆s = 3.53.

3.1.2 Amplitude Thresholds

Amplitude thresholds are used to limit the number of peaks in each analysis frame. The

desire is to detect and track only the most perceptually significant partials. Because

many sounds of interest exhibit a high frequency roll-off, it is helpful to equalize the

spectrum. This results in an analysis that gives equal weight to both high and low

frequency components and avoids resynthesis that sounds dull and heavily lowpass

filtered.

Page 55: Klingbeil Dissertation Web

3. Sinusoidal Analysis and Resynthesis 45

Equalization can be achieved by applying a pre-emphasis high-pass filter in the time

domain (Maher 1989), or by applying an emphasis curve to the STFT data (Serra 1989).

One particular disadvantage of the applying a filter in the time domain is an alteration

of the magnitude and phase spectra, the effects of which can only be easily undone by

applying a complementary de-emphasis filter following resynthesis.

SPEAR achieves the effect of pre-emphasis by using a frequency dependent amplitude

threshold curve. SPEAR establishes two amplitude thresholds: Tb, which is a variable

threshold that must be exceeded to begin a new partial track, and Td, which is fixed lowest

threshold (default value is −90 dB). If peaks are found that exceed Td, then they may be

used to continue an existing partial track. If no appropriate track continuation is found,

the partial ends — so Td may be considered a “death” threshold for possible continuation

of a partial.

Tb is computed using a frequency dependent threshold curve, At( fp), in combination

with the maximum bin amplitude of the analysis frame, amaxn . Incorporating the maximum

bin amplitude allows Tb to track the overall signal level. The threshold curve is given by

At( fp) = aT + aL +(

aR

b− 1

)−(

aR

b− 1

)b fp/20000 (3.2)

where fp is the frequency in hertz, b is a parameter controlling the shape of the curve, aT

is user controllable threshold in dB, aL is the amplitude offset at 0 kHz in dB, and aR is

the range of the curve in positive dB from 0 to 20 kHz. Figure 3.3 shows At( fk) with the

default values b = 0.0075, aT = −60, aL = 26, and aR = 32.

Given the maximum bin amplitude in dB for a frame, amaxn and the frequency fp in Hz

of the peak under consideration, the threshold Tb is given by

Tb = amaxn + At( fp) (3.3)

Tb constitutes a “birth” threshold for potential new partials. The birth threshold requires

low frequency peaks to exceed a greater threshold than high frequency ones. In addition

to helping find more high frequency partial tracks, this aids in the rejection of spurious low

frequency, low amplitude partial tracks. Although the high frequency emphasis afforded

Page 56: Klingbeil Dissertation Web

3. Sinusoidal Analysis and Resynthesis 46

Figure 3.3: Frequency dependent threshold curve with the default values b = 0.0075,aT = −60, aL = 26, and aR = 32.

by the variable threshold provides increased resynthesis fidelity for many real world

sounds, it can result in many low amplitude, high frequency partials. In many cases, the

signal energy above 5 kHz could be more efficiently and robustly modeled as noise bands,

rather than clusters of low amplitude sinusoids. Possible approaches to modeling noise

will be discussed in section 3.6.

3.2 Partial Tracking Using Linear Prediction

Peaks that are retained following thresholding are joined into breakpoint functions rep-

resenting individual sinusoidal partials. The linear prediction (LP) method of Lagrange

et al. (2003) is used to connect the peaks into tracks. Linear prediction treats each evolving

sinusoidal track as a signal. The current signal sample, x(n), is approximated as a linear

combination of past samples:

x(n) =K

∑k=1

a(k)x(n− k) (3.4)

There are K linear prediction coefficients, a(k), that are calculated by minimizing the error

between the predicted value, x(n), and the actual value x(n). The K coefficients can then

be used to predict successive values x(n + 1), x(n + 2), etc. Several different algorithms,

Page 57: Klingbeil Dissertation Web

3. Sinusoidal Analysis and Resynthesis 47

including the Burg method, autocorrelation, and covariance can be used to compute a(k).

Lagrange et al. show that the Burg method is most suitable in this application.

LP can be used to compute possible future values for both frequency and amplitude.

The frequencies and amplitudes of newly detected peaks are compared to the predicted

values, and the closest matches are used to extend the sinusoidal tracks. The method is

implemented as follows: at frame k we have detected N candidate peaks and we have

M active sinusoidal tracks which extend to at most frame k − 1. For each of the M

active sinusoidal tracks, the Burg method is used to compute two sets of linear prediction

coefficients, one on frequency and one on amplitude. An LP model of order 6 computed

from a maximum of 64 previous values has worked well in practice. It is critical that

enough points are used to capture periodic features such as amplitude modulation or

vibrato. For a more thorough discussion of the choice of LP parameters, see Lagrange et al.

(2007).

The LP coefficients for track m are used to predict frequency f prm and amplitude apr

m

values for frame k. When a new sinusoidal track starts and there are not enough values to

compute the LP coefficients, the mean frequency and amplitude are used as the predicted

values. The error between predicted values for track m and peak n are given by a measure

of Euclidean distance

Em,n =

√[12 lg

(f obsn

f prm

)]2

+[

α 20 log10

(aobs

n

aprm

)]2

(3.5)

where f obsn and aobs

n are the observed frequency and amplitude values for peak n. The first

term measures distance in semitones and the second in dB with a scale factor α. Informal

tests have shown α = 112 to offer a reasonable weighting between prediction errors in

frequency versus those in amplitude.

Each track m selects the best continuation peak n over all N peaks subject to constraints

on the maximum allowable difference in predicted versus actual frequency. More precisely,

| f obsn − f pr

n | < ∆ fmax where ∆ fmax is proportional to the analysis frequency fa. For harmonic

sounds, ∆ fmax = 34 fa is a reasonable default value. ∆ fmax is a user-controllable parameter

Page 58: Klingbeil Dissertation Web

3. Sinusoidal Analysis and Resynthesis 48

and can be reduced for sounds with very stable partials or increased for sounds with

partials that exhibit wide frequency excursions.

As with guide based tracking methods, linear prediction partial tracking can span

temporal gaps in the analysis data. Tracks that fail to find a match either become inactive

or lie dormant for several successive frames. In the case of a track that has been dormant

for j frames, linear prediction is used to predict the j + 1th values for matching to candidate

peaks.

3.3 Resynthesis

SPEAR offers several different synthesis methods. For real-time playback, the inverse

FFT method (abbreviated as IFFT or FFT−1) offers excellent efficiency and very good

quality (Rodet and Depalle 1992). Although most synthesis artifacts of the standard IFFT

method can be minimized with the use of appropriate overlap-add windows, extremely

rapid modulations of frequency or amplitude may not be adequately reproduced. For the

highest quality sound, SPEAR can perform oscillator bank synthesis with optional cubic

phase interpolation as in the classical McAulay-Quatieri method. SPEAR also supports the

resynthesis of Loris RBEP files which include a noise component for each breakpoint (Fitz

1999). These so-called bandwidth enhanced partials can be synthesized with either the

IFFT method or noise modulated oscillators.

3.3.1 Inverse FFT Synthesis

Since the DFT is invertible, it is possible to synthesize any desired time domain waveform

by constructing the complex spectrum of the desired signal and applying the inverse DFT.

In this case we are interested in the synthesis of sinusoids which, as detailed in section

2.1.4, have well understood spectra. A time domain windowed sinusoid can be synthesized

by the following procedure:

1. shift the complex spectrum of the window function so that it is centered on the bin

frequency of the desired sinusoidal frequency

Page 59: Klingbeil Dissertation Web

3. Sinusoidal Analysis and Resynthesis 49

2. scale the spectrum of the window function according to the desired complex ampli-

tude (magnitude and phase)

3. accumulate the scaled shifted window function into the DFT buffer

4. perform the inverse DFT

Steps 1–3 correspond to convolving the window spectrum with the spectrum of a sinusoid

(a line spectrum). Because convolution is a linear operation, steps 1–3 can be performed

for each sinusoid to be synthesized. The inverse DFT can then computed as a final step.

The final DFT only adds a constant amount of work regardless of the number of sinusoids.

Provided that steps 1–3 can be carried out efficiently, we have the basis for an algorithm to

quickly synthesize a sum of many sinusoids.

Step 1 can be performed efficiently by pre-computing an oversampled spectrum of the

window function. For a DFT/IDFT of length N, we first compute a time domain window

function of h(n) of length N. The window function is then zero padded (zeros appended)

so that it has length M = N × R, where R is the oversampling factor (R is typically chosen

to be a power of 2). Call this zero padded window h′(n).

h′(n) =

h(n) for 0 ≤ n < N

0 for n < 0 or n ≥ N(3.6)

The DFT of the zero padded window gives the oversampled interpolated spectrum of the

window H(k).

H(k) =N−1

∑n=0

h′(n)e−i2π kN n k = 0, 1, 2, . . . , N − 1 (3.7)

The window spectrum can then be shifted to any fractional bin position that is a multiple

of1R

. For bin frequencies that are not a multiple of1R

, the window value may be

approximated by index truncation, index rounding, linear interpolation, or some higher

order interpolation scheme.

Steps 2–3 — scaling and accumulating the complex spectrum of the window func-

tion — can be performed with various degrees of efficiency depending on the precision

required. To precisely synthesize the desired sinusoid, all N values of the window spec-

Page 60: Klingbeil Dissertation Web

3. Sinusoidal Analysis and Resynthesis 50

trum must be scaled and accumulated into the DFT buffer. Equation 3.8 summarizes

the shifting, scaling, and accumulation of the window spectrum W(n) for a sinusoid of

amplitude An and bin frequency of bn. Note that to synthesize a real valued-sinusoid,

we must represent half of the amplitude as a complex sinusoid, and half as its complex

conjugate.

X(k) =An

2H(b(k− bn)Rc)eiφ +

An

2H(b(N − k− bn)Rc)e−iφ (3.8)

The symmetry of the DFT buffer means that X(k) can be constructed by first accumulating

N/2 values in the positive frequency half (0 ≤ n ≤ N/2), and then reflecting their complex

conjugate in the negative frequency half (N > n > N/2).

The procedure implies one table lookup (for the window function), and one complex

multiply (four scaler multiplies) for each of the N2 + 1 values. Compared to synthesis via a

table lookup oscillator, this method is actually less efficient. The oscillator requires one

table lookup and one scaler multiply for each of the N samples. However, significant

savings can be achieved if we approximate the scaled and shifted window function

spectrum.

Although the complete window function spectrum has length N, it should be noted

that many of the values are close to zero. Therefore, we can approximate steps 2–3 by

scaling and shifting only J window samples immediately surrounding the main lobe

(typically between 7–15 samples). Laroche (2000) calls the restricted set of samples the

“spectral motif.” Accumulating only the spectral motif (rather than the entire window

spectrum) reduces the number of table lookups and complex multiplies to a constant J,

regardless of the DFT buffer size.

Although any window function can be used, for computational efficiency we desire a

window function with a reasonably narrow mainlobe and low sidelobes. When the DFT

buffer is filled, only the values of the main lobe and the first few side lobes need to be

computed.

The result of the inverse DFT is a sum of sinusoids (each with constant amplitude

and frequency) that have been multiplied by the time domain window function h(n).

Page 61: Klingbeil Dissertation Web

3. Sinusoidal Analysis and Resynthesis 51

To synthesize time varying amplitudes and frequencies, an overlap-add process can be

used. We begin by multiplying each sample in the DFT buffer by 1h(n) . This effectively

removes the time domain effect of the window, leaving a constant amplitude sinusoid. We

then apply a triangular synthesis window s(n) which gives the desired linear amplitude

interpolation when overlap-added together.

Since frequency is constant across each DFT buffer, there will be some modulation

artifacts resulting from overlap-adding. These effects can be reduced by carefully matching

the phase of each sinusoid from one buffer to the next. For an overlap factor of 2, the

phase should be matched at the midpoint of the triangular slope (Rodet and Depalle 1992).

An important implementation detail of the IFFT method concerns the imprecision

introduced by the approximation of the window spectrum. By only using J samples

centered on the main lobe, the window spectrum has been band-limited. Effectively the

window spectrum has been multiplied by a rectangular window. Since multiplication in

the frequency domain results in convolution in the time domain, the time domain window

is no longer limited to N samples. The result is time domain aliasing which introduces

distortion in the resynthesized sinusoid. This distortion is particularly evident at the edges

of the time domain buffer where the window function approaches its minimum. When

the window function is divided out (multiplied by 1h(n) ), these distortions are dramatically

amplified. It is particularly evident for window functions that taper to zero at the edges

(windows in the Blackman-Harris family, for example). Figure 3.4 compares a sinusoid

with IFFT synthesized versions using different values of J. Note that even with large values

of J, significant error is still present. To avoid this problem, some number of samples D at

the beginning and end of the DFT buffer must be discarded. D is typically on the order ofN8

, so the reduction in synthesis efficiency is negligible.

Inverse FFT synthesis appears to offer some open areas for continued research. More

detailed studies of signal distortion in inverse FFT synthesis would be welcome. It would

be helpful to have a more precise analysis of the effects of different synthesis windows,

window oversampling and interpolation strategies, the choice of J (controlling the size

Page 62: Klingbeil Dissertation Web

3. Sinusoidal Analysis and Resynthesis 52

Figure 3.4: Comparison of sinusoids generated via IFFT synthesis with different values ofJ (the spectral motif length) showing the first 16 samples of the time domain signal whereN = 64, frequency = 1.5 cycles per N samples. Window type is Blackman-Harris.

of the spectral motif), and the effect of D on signal to noise ratio. Jean Laroche’s (2000)

investigations provide some important preliminary insights in this area.

3.3.2 IFFT Synthesis Implementation

With an overlap add factor of 2, each inverse DFT computesN − 2D

2new output samples.

In the current implementation, SPEAR performs IFFT synthesis with N = 512, D =N8

.

These parameters allowed for real-time resynthesis of hundreds of partials on relatively

modest hardware.1 Future versions of SPEAR may allow the user to fine tune the resyn-

thesis parameters. In particular, it may be desirable to reduce N. Because frequency and

amplitude are constant across each DFT buffer, rapid frequency changes may not be well

represented with larger N.

In the current implementation, the spectral motif length, J, is fixed at 9. Because of

the problem of distortion at edge of the DFT buffer, the Hamming window, as defined

in equation 3.9, was chosen for the synthesis window. Since we divide by the window

function, windows that taper to zero will amplify distortions rather dramatically as the

1PowerPC G3 class processor at 400 MHz — circa 2003

Page 63: Klingbeil Dissertation Web

3. Sinusoidal Analysis and Resynthesis 53

inverse tends toward infinity. The minimum value of the Hamming window at the buffer

edge is2

23.

h(n) =

2546− 21

46cos

(2πn

N − 1

)for 0 ≤ n < N

0 for n < 0 or n ≥ N

(3.9)

Real time resynthesis proceeds according to a time counter t that indicates the current

location in the analysis data. Whenever a new output buffer is required, the synthesis

callback checks to value of t to determine the set of sinusoids that need to be synthesized.

The center of each DFT synthesis window is represented by t. Pn is the set of partials

active at time tn. The amplitude and frequency for each partial in Pn is determined by

linear interpolation. At time tn+1, we have a new set of partials Pn+1. All partials in

Pn⋂

Pn+1 must have phase continuity at the overlap-add midpoint. This is achieved by

maintaining a set of virtual oscillators that store the previous phases. Each oscillator

structure is associated with an active partial by means of a hash table (Cormen et al. 1990).

Thus, it is reasonably efficient to lookup phase as partials turn on and off. Moreover t

can proceed at any rate through the analysis data. This facilitates the ability to manually

“scrub“ forward and backward through the sound. The analysis data can also be freely

transformed concurrent with resynthesis. SPEAR does appropriate data locking to make

sure the analysis data structures remain consistent between the concurrent resynthesis and

user interface threads.

3.3.3 Oscillator Bank Resynthesis

SPEAR can optionally perform a non-realtime oscillator bank resynthesis with linear

interpolation of frequency and amplitude as shown in equation 2.24 (page 40). This

provides a much higher fidelity result than the IFFT method.

Page 64: Klingbeil Dissertation Web

3. Sinusoidal Analysis and Resynthesis 54

3.3.4 Phase Management

As noted in section 2.4.4, linear interpolation of frequency will, in general, only retain

the phase of the first breakpoint of a partial. Although in many situations the phase

information may not be perceptually important, there are cases where temporal phase

coherence between partials is quite important, for example in the reproduction of transients.

We also seem to be particularly attuned to the phase relationships in speech signals.

McAulay and Quatieri (1986) developed a frequency/phase interpolation method that

maintains phase and frequency at each breakpoint. Consider two successive breakpoints i

and i + 1 with phases φi and φi+1 and frequencies ωi and ωi+1 (here we are expressing

frequency in radians per sample rather than cycles per second). Since frequency is the

derivative of phase, we desire a phase interpolation function with endpoints equal to φi

and φi+1 and first derivatives at the endpoints equal to ωi and ωi+1. Independent control

of the slope at two points requires a cubic function, so we begin by defining a cubic phase

of the form

θ(n) = ζ + γn + αn2 + βn3 for n = 0 . . . S (3.10)

where n is the sample index and S is the number of samples from breakpoint i to breakpoint

i + 1. The first derivative (the frequency function) is

θ′(n) = γ + 2αn + 3βn2 for n = 0 . . . S (3.11)

The endpoints at sample 0 and sample S are

θ(0) = ζ = φi (3.12)

θ′(0) = γ = ωi (3.13)

θ(S) = ζ + γS + αS2 + βS3 = φi+1 (3.14)

θ′(S) = γ + 2αS + 3βS2 = ωi+1 (3.15)

With four equations we can solve for the four unknowns. ζ is equal to the initial phase, and

γ is equal to the initial frequency, which leaves only to solve for α and β. One complication

Page 65: Klingbeil Dissertation Web

3. Sinusoidal Analysis and Resynthesis 55

π

π

0

θ(n) = φi + ωin + α(M)n2 + β(M)n3

φi

slope = ωislope = ωi+1

φi+1

φi+1 + 8π M = 4

φi+1 + 6π M = 3

φi+1 + 4π M = 2

φi+1 + 2π M = 1

M = 0

n = 0 n = S

sample index

phas

e

Figure 3.5: Family of cubic phase interpolation functions with different values of M (afterMcAulay and Quatieri (1986)).

is that the final phase is measured modulo 2π, so in fact θ(S) should be expressed as

θ(S) = ζ + γS + αS2 + βS3 = φi+1 + 2πM (3.16)

for some integer M. M defines a family of cubic phase functions with different amounts

of phase unwrapping (see figure 3.5).

The solutions for α and β are parameterized in terms of M as follows:

α(M) =3S2 (φi+1 − φi −ωiS + 2πM)− 1

S(ωi+1 −ωi) (3.17)

β(M) =−2S2 (φi+1 − φi −ωiS + 2πM) +

1S2 (ωi+1 −ωi) (3.18)

McAulay and Quatieri show that the optional choice of M is that which makes the phase

curve maximally smooth (the maximally smooth choice for figure 3.5 is M = 2). This can

be defined in terms of minimizing the second derivative of θ(n) with respect to sample

Page 66: Klingbeil Dissertation Web

3. Sinusoidal Analysis and Resynthesis 56

index. The result is that the maximally smooth M is the closest integer to x, where

x =1

[(φi + ωiS− φi+1) + (ωi+1 −ωi)

S2

](3.19)

The complete phase function is

θ(n) = φi + ωin + α(M)n2 + β(M)n3 for n = 0 . . . S (3.20)

It is important to note that cubic phase interpolation will generally only work for unmodified

resynthesis. Partials that have undergone time dilation, transposition, or frequency shifting

cannot be synthesized reliably with this method since the resulting phase functions are no

longer “smooth” enough to avoid frequency modulation artifacts. Methods for preserving

phase under modifications remain an interesting open problem. It may prove effective to

resynchronize phase only at certain moments (as suggested by the work in Robel (2003)

and Fitz and Haken (2002)), or slightly modify the breakpoint frequencies to smooth out

the phase function. Phase interpolation in the unmodified case is important for recovering

the residual signal, which is discussed in section 3.6.

3.4 Data Storage

A popular data storage model in analysis-synthesis applications is a sorted list of time

frames. Each frame contains a list of peaks (amplitude, frequency, and phase) that are

connected to successive time frames via either index numbers or pointers.

Rather than represent the analysis data as a sorted list of time frames, SPEAR uses

a list of partials that are represented by breakpoint functions of time versus amplitude,

frequency, and phase. With this storage model the implementation of cut, copy, and paste

operations, of both entire partials or segments, is straightforward. Additionally, partials

can be individually shifted, stretched, or compressed in time without resampling to fixed

frame time points. A further advantage is that the storage model can easily support

multirate analysis data (Levine 1998) or time reassigned breakpoints (Fitz 1999).

Page 67: Klingbeil Dissertation Web

3. Sinusoidal Analysis and Resynthesis 57

3.4.1 Time-span Frames

A common operation, particularly during synthesis, is determining the list of partials (and

breakpoint segments) crossing a particular time point. In the frame based storage model,

a binary search of sorted frames quickly determines the partials that are active at any

particular time.

With a list of breakpoint partials, this query requires an iteration through all the

partials in the data set. For a short sound, this may only require looking at several hundred

elements, but for longer sounds there are likely to be many partials of short duration.

For example, the total number of partials for a one minute sound could be well over

twenty-thousand. During synthesis, continued iteration over thousands of partials is too

inefficient. A solution is to maintain a parallel data structure of time-span frames. For

each frame, a list is kept of all partials that have breakpoints within the frame’s time

span — typically on the order of 0.125 seconds. Figure 3.6 shows an example of ten partials

segmented into five frames where each frame contains a list of its active partials. The

active partial lists maintain a relatively constant size as partials turn on and off. For a

typical sound, this reduces active partial lookups to an iteration over only several hundred

elements.

0

1

2

3

4

5

6

7 8

9

t0 t1 t2 t3 t4 t50, 1, 2,

3, 40, 1, 2, 3, 4, 5

0, 1, 3,4, 6, 7

0, 1, 3, 8 0, 1, 8, 9

Figure 3.6: Data structure for division of partials into time-span frames. Comma delimitedlists for each frame indicate the active partials.

Page 68: Klingbeil Dissertation Web

3. Sinusoidal Analysis and Resynthesis 58

3.5 Transient Sharpening

As observed in section 3.1.1, the choice of window length directly affects both the frequency

and temporal resolution of the analysis. For polyphonic or inharmonic sounds in particular,

we desire a long analysis window so as to be able to resolve closely spaced partials.

However, as the window length increases, the temporal resolution decreases. This is a

particular problem for sounds that contain transients. All signal energy is spread across

the window and assigned to a breakpoint located at the temporal center of the analysis

window rather than at the precise location of the transient. Moreover, since analysis

windows overlap, the energy will be distributed across multiple windows and assigned to

multiple breakpoints. This results in the familiar pre- and post-echo effects of traditional

STFT analysis/synthesis.

Fitz (1999) shows how the method of reassignment can be used to improve the temporal

resolution and reduce pre- and post-echo. The method of reassignment begins by noting

that traditional STFT analysis assigns all energy included in the window to the temporal

center of the window. When there is significant energy in a particular frequency band that

is located near the edge of a window, this energy should instead be assigned to the center

of gravity (time-frequency centroid) of the window. This time-frequency coordinate can be

computed from the derivatives of the phase spectrum. The time-frequency centroid will

be located at the point where phase is changing most slowly (where the partial derivative

of phase with respect to time and phase with respect to frequency is zero).

Auger and Flandrin (1995) and Plante et al. (1998) show how to efficiently compute the

point of reassignment as a ratio of Fourier transforms. The reassigned time for frequency

bin k is given by

t(k) = t +<e

{Xht(k) Xh(k)|Xh(k)|2

}(3.21)

where t is the original center of the analysis window (in samples). Xh(k) is the DFT of

the signal using the usual analysis window h(n). Xht(k) is the DFT of the signal using a

special time-weighted analysis window ht(n). ht(n) is defined as the product of h(n) and

Page 69: Klingbeil Dissertation Web

3. Sinusoidal Analysis and Resynthesis 59

a ramp function from −M−12 to M−1

2 , where M is the window length.

ht(n) = h(n)[

n− M− 12

]n = 0, 1, 2, . . . , M− 1 (3.22)

The real portion of this ratio represents the time reassignment in samples.

The reassigned frequency for bin k is given by

ω(k) = ωk −=m

{Xhω(k) Xh(k)|Xw(k)|2

}(3.23)

where ωk is the bin frequency, and Xhω(k) is the DFT of the signal using a frequency-

weighted analysis window hω(n). hω(n) can be computed by taking the DFT of h(n),

multiplying by a frequency ramp, and then taking the inverse DFT. Note that the time-

frequency center of gravity is computed independently for each frequency bin.2

Pre- and post-echo can be dramatically reduced by removing breakpoints with large

time corrections (Fitz 1999; Fitz and Haken 2002). Specifically, if the time correction is

larger than analysis hop size (and the hop size is less than the window length — which is

always the case in this implementation), a breakpoint can be safely removed since it will

be detected in a neighboring frame. Figures 3.7 and 3.8 show the analysis of a castanet

click that occurs at about 0.868 seconds.

Note that the transient event is well localized, and pre- and post-echo are dramatically

reduced. In order to avoid a complete amplitude discontinuity at the onset of a partial,

a zero amplitude breakpoint is joined to the head of each partial. Informal observations

indicate that although the reassignment method does an excellent job of localizing sharp

transients, there is some loss of energy, likely due to the removal of the breakpoints with

large time corrections. The addition of a zero amplitude breakpoint creates a short fade

into each transient which has the additional benefit of raising the average signal level

around the transient event. Further experimentation should be conducted to determine

whether transient breakpoints would benefit from an additional amplitude boost.

2 Note that equations 3.21 and 3.23 differ from those in Auger and Flandrin (1995), Plante et al. (1998), andFitz (1999) in that we add the reassignment term for time reassignment and subtract for frequency reassignment.The sign depends on whether or not the STFT is defined with a time reversed window. We use the definition in

Page 70: Klingbeil Dissertation Web

3. Sinusoidal Analysis and Resynthesis 60

Figure 3.7: Breakpoints from the analysis of a castanet click using the standard sinusoidalmodel.

Figure 3.8: Breakpoints from the analysis of a castanet click using the time reassignmentmethod. Breakpoints with large time corrections have been removed.

Page 71: Klingbeil Dissertation Web

3. Sinusoidal Analysis and Resynthesis 61

3.6 Noise Modeling

Serra and Smith were the first to develop extensions to sinusoidal modeling for representing

noise. The SMS (Spectral Modeling Synthesis) system offers two different ways to model

non-sinusoidal components. In the residual model, the sound is first analyzed using

sinusoidal modeling. Next, the signal is resynthesized using the oscillator bank method

with cubic phase interpolation as described in section 3.3.4:

x(n) =N−1

∑i=0

Ai(n)cos(θi(n)) (3.24)

The residual signal e(n) is then computed by subtracting the resynthesized signal x(n)

from the original x(n).

e(n) = x(n)− x(n) (3.25)

Because phase is preserved in the resynthesis, this subtraction can take place in the time

domain. The residual e(n) will typically consist of noise, attack transients, and pre- and

post-echo. The original signal can of course be perfectly reconstructed by adding the

residual to the resynthesis. However, typically we wish to make some sonic transformations

prior to resynthesis. In this case a more flexible modeling of the residual is required.

The stochastic model represents the residual in the frequency domain. For each frame l

of the analysis, we compute a residual magnitude spectrum from the magnitude spectrum

of the original signal and the magnitude spectrum of the phase accurate resynthesis.

|El(k)| = |Xl(k)| − |Xl(k)| (3.26)

Xl(k) is the DFT of the original signal at frame l and Xl(k) is the DFT of the sinusoidal

resynthesis. Assuming that e(n) is a stochastic signal, it can be modeled as a set of

magnitude spectra with random phase. Time stretching or transposition of the residual

can be accomplished by interpolating new magnitude spectra |El(k)|, setting the phase

equation 2.13 (page 20) in which the window is not time reversed. Auger and Flandrin and Plante et al. definetheir STFT with a time reversed window. Thanks to Kelly Fitz for explaining the reason for this difference.

Page 72: Klingbeil Dissertation Web

3. Sinusoidal Analysis and Resynthesis 62

spectra to random values between −π and π, applying the inverse DFT, and then overlap

adding.

Optionally |El(k)| can be computed directly in the frequency domain by subtracting the

frequency domain peaks detected in frame l from |Xl(k)|. |El(k)|may also be approximated

as a spectral envelope. The spectral envelope can be viewed as a filter that shapes a flat

noise spectrum.

Several variants of the deterministic plus stochastic model have been proposed. Pampin

(2004) developed the ATS system which separates the energy in El(k) into 25 bands; each

band corresponding to a step on the Bark frequency scale. The noise energy in each band

may then be optionally redistributed among the partials.

3.6.1 Bandwidth Enhanced Sinusoids

This approach is similar to the bandwidth enhanced additive sound model described in

Fitz (1999) and implemented in Loris. Bandwidth enhanced synthesis models the signal

as a sum of bandwidth enhanced oscillators — sinusoidal oscillators that are amplitude

modulated by a lowpass filtered noise signal χ(n).

y(n) = [A(n) + β(n)χ(n)] cos(

n2πF(n)

fs

)(3.27)

Equation 3.27 defines a bandwidth enhanced oscillator with time varying amplitude A(n)

and time varying frequency F(n). The amplitude of the lowpass filtered noise is controlled

by β(n). Multiplication of a sinusoid by noise in the time domain results in the convolution

of the noise spectrum with the sinusoid spectrum. The multiplication shifts the noise

spectrum so that it is centered at frequency F(n).

In Loris, each breakpoint has an additional noise parameter κ that varies between 0

and 1. If A(n) is the local average partial energy, the the bandwidth enhanced oscillator is

defined by

y(n) = A(n)[√

1− κ +√

2κχ(n)]

cos(

n2πF(n)

fs

)(3.28)

Page 73: Klingbeil Dissertation Web

3. Sinusoidal Analysis and Resynthesis 63

When κ = 0, y(n) is a pure sinusoid. As κ increases, the amplitude of the narrowband

noise also increases. This model has the advantage that noise is tightly coupled to the

sinusoidal partials. As the sound is edited and manipulated — for example as partials are

transposed — the noise will follow. It is sometimes convenient that the noise amplitude is

controlled by a single parameter κ. However since the noise parameter is coupled to overall

partial amplitude, several square root operations must be invoked, which is somewhat

inefficient for realtime resynthesis.

SPEAR currently does not implement noise analysis, but it will resynthesize noise of

partials imported from Loris SDIF files (see section 4.6 for more information about the

use of SDIF). Loris defines the lowpass filtered noise signal as white noise filtered by a

3rd order Chebychev IIR filter with cutoff of 500 Hz, and SPEAR attempts to match this

behavior. Although the Loris noise representation is compact (based on a single parameter

κ), it lacks generality and is not an ideal choice for cross program data exchange. A more

flexible system would specify noise amplitude and bandwidth independently of sinusoidal

amplitude. Such a data model would better support different noise analysis strategies and

would allow for experimentation with different noise bandwidths. Expanding the number

of noise parameters would require either a new SDIF matrix type or an extension of an

existing type. Expanding the noise handling capabilities of SPEAR is an important area

for future work.

Page 74: Klingbeil Dissertation Web

64

4. Editing and Transformation

4.1 Graphical User Interface

One of the motivations behind the development of SPEAR was the desire for a graphical

interface that would allow the user to make a quick visual assessment of analysis quality.

In spite of the best efforts in the choice and implementation of the analysis algorithms,

sonic artifacts are common. Even with a reduced set of available analysis parameters,

some trial and error is usually necessary to determine the optimal settings for a particular

application.

With systems that are command-line oriented (Loris, ATS, SNDAN), the user is left

to either resynthesize and listen carefully for artifacts, or use an additional plotting or

visualization tool. SPEAR offers an intuitive interface that allows the user to quickly zoom

in on analysis artifacts. Problematic areas of the sound can be isolated and auditioned,

and the data can be modified as desired. For example, spurious sinusoidal tracks can be

deleted, attenuated, or joined with other neighboring tracks.

The analysis data is displayed with time on the abscissa and frequency on the ordinate.

As in a grayscale spectrogram, varying levels of shading indicate the relative amplitudes

of partials. Figure 4.1 illustrates the overall user interface. Following the model of time

domain waveform editors, the amount of detail shown for the breakpoint functions varies

according to a user controlled zoom factor. At high zoom levels, every breakpoint is shown

and clickable handles allow individual manipulation in time and frequency (figure 4.2).

At lower zoom levels, the decrease in detail avoids visual clutter and significantly speeds

redraw.

Page 75: Klingbeil Dissertation Web

4. Editing and Transformation 65

Figure 4.1: Zoomed-out display showing lasso selection and playback control sliders.

Figure 4.2: Zoomed-in display showing amplitude adjustment dialog.

Page 76: Klingbeil Dissertation Web

4. Editing and Transformation 66

scroll

select frequency region

lasso selection

transpose selection

time shift selection

playback scrub

draw partial

time stretch/compress

time/frequency shift

frequency shift

rectangular selection

point/partial selection

selection time region

Figure 4.3: The SPEAR tool palette.

A tool palette (figure 4.3) offers different modes for graphical selection and manipula-

tion. Interaction follows a selection and direct manipulation paradigm (Shneiderman 1983).

For example, clicking on a partial selects it and shift-clicking adds partials to the current

selection. Choosing the transpose tool then allows the selected partials to be transposed by

dragging up or down. Additional selection modes allow the user to sweep out areas in

time and frequency or to draw arbitrary time-frequency regions with a “lasso.” Following

the model of graphics editors, arbitrary selections can be made up of a union or difference

of individual contiguous selections.

Editing tools and commands that operate on the current selection include transposition,

frequency shifting, time shifting, time stretching, and amplitude adjustment. Unlimited

undo/redo is supported for all editing operations, and most editing can be performed

while the sound is being synthesized. For further real-time control, sliders allow adjustment

of the overall transposition level, amplitude, and playback speed (figure 4.4). New partials

can be added by drawing with the pencil tool. New breakpoints are added with amplitude

proportional to the dragging speed.

Figure 4.4: The SPEAR playback control sliders.

Page 77: Klingbeil Dissertation Web

4. Editing and Transformation 67

4.2 Selection Model

During the development of SPEAR, two possible selection models and implementations

were considered. The first may be termed the “region selection“ method. In this model, a

selection is represented as a polygon (concave or convex) in the time-frequency plane. (It

would also be possible to represent the selection in the time-amplitude plane if desired.)

Selections would be defined by sweeping out regions graphically or by setting numeric

parameters. More complex selections could be represented as the union of several regions.

The second selection model is “point selection“ method. In this case, the selection is

represented as a collection of references to the specific breakpoints of interest.

Given the desire to allow both extremely detailed editing and complete flexibility in

the temporal positioning of breakpoints, it was decided that the point selection model was

best. To reduce implementation and interface complexity, this is the only selection model

that has been implemented to date.

There are a few drawbacks to this model. In some cases it may be desirable to define a

transformation (such as an amplitude envelope) that occurs across a certain time span. In

this case, the region selection model is ideal. When the user requests such an operation

on a point selection (for example, an amplitude fade), the bounding region of the point

selection must first be computed. It may also be desirable to define a time-frequency

region (such as a formant) that remains fixed when partials are shifted or transposed. In

this case, the region selection model is required.

Point selections are implemented by associating a selection list with each breakpoint

array. The selection is stored efficiently as a range set data structure (Phillips 1999). Rather

than maintaining the index of each breakpoint, the range set represents contiguous runs

of breakpoints by the start and ending points of the run. More formally, if n breakpoints

starting at index i are selected, the range set represents this selection as the pair (i, i + n),

rather than the list [i, i + 1, i + 2, . . . , i + n− 1]. Discontiguous selections are represented as

an ordered list of pairs:[(i0, i0 + n0), (i1, i1 + n1), . . . , (ij, ij + nj)

]. Given that contiguous

selections are the most common case, the range set is very efficient in both time and

Page 78: Klingbeil Dissertation Web

4. Editing and Transformation 68

space requirements. The range set also supports operations to efficiently add and coalesce

ranges, as well as delete and split ranges.

4.2.1 Rule-based Selection

Selections can also be made programatically. For example, the user can choose to select all

partials with a duration less than some specified value. Or, the user can choose to select

partials with an average amplitude less than some desired level. Currently the number of

choices for rule-based selection is quite limited, however one can imagine any number of

selection criteria: minimum or maximum frequency, average frequency, frequency range,

frequency periodicity (vibrato rate), phase stability, amplitude range, maximum amplitude,

etc. Comparison (=, >, <) and logical operations (NOT, AND, OR, XOR) could be used to

develop any number of rule-based selection criteria. The extent of possibilities suggests

that a scripting language (rather than a GUI) would be best way to implement flexible

rule-based selection. It is hoped that future versions of SPEAR will include this capability.

4.3 Editing Features

The SPEAR document model supports cut, copy, and paste of any selected data. When

pasting data from one document to another, care must be taken to make sure data types

match. Breakpoints are required to have time, frequency, and amplitude components. They

may also have phase and noise components, but these are optional (for a discussion of

the noise component, see section 3.6.1, page 62). Analyses performed with SPEAR will

have at least time, frequency, amplitude, and phase. Since analysis data may be imported

from other software and via different file formats (see section 4.6), the document model

must be prepared to support different breakpoint types. When pasting into an existing

document, the pasted data is modified to match the destination document’s breakpoint

type. For example, when pasting data that has only time, frequency, and amplitude into a

document that also has phase, the newly pasted data will have phases set to zero.

Page 79: Klingbeil Dissertation Web

4. Editing and Transformation 69

In future versions of SPEAR it may prove beneficial to allow greater flexibility of point

types. Additional data types could specify specify stereo panning, 3D spatial location,

noise bandwidth, etc. Partials are currently implemented as multi-dimensional arrays

with contiguous allocation. A flexible point model would be better implemented with

collections of one-dimensional arrays, each array corresponding to one data type.

contiguous multi-dimensional array

collectionof

arrays

Figure 4.5: Different breakpoint storage models.

SPEAR supports unlimited undo/redo for all editing operations. Early in the design

process it was decided that the undo implementation should be simple and robust, yet

flexible enough to support arbitrary edits and transformations. Each undo/redo operation

is modeled as an ordered collection of sub-operations consisting of any of the following:

• selection

• edit partial

• insert partial

• remove partial

These sub-operations are bundled together into a single undo or redo operation. An

undo stack stores the state before the operation (this is the data required to restore the

document to its previous state), while the redo stack stores state after the operation. To

avoid excessive dynamic memory use, the data for edit, insert, and remove operations is

stored on disk. Selections are stored in locally addressable memory. Generally this does

not present a storage problem since the range map data structure is quite compact.

Page 80: Klingbeil Dissertation Web

4. Editing and Transformation 70

A selection operation consists of the partial indices and breakpoint selection ranges.

An edit operation is defined as any operation on a partial that does not require the

insertion or removal of any other partials (and therefore does not change any partial index

numbers). Insert or remove operations are those that will alter the partial indices in the

data model. Operations that can split partials into multiple segments (such as cut or delete)

are modeled as a sequence of edit, insert, and remove operations. It is crucial that the

atomic sub-operations are ordered properly so that the precise order of partial indices is

always restored.

4.4 Time Expansion and Contraction

SPEAR supports two different modes for time expansion and contraction. The first is

the “independent“ mode. In this mode, each partial is expanded or contracted relative

to the starting time of the partial (or selection), which remains fixed in time. If tk(n)

represents the time of nth breakpoint for partial k, then the independent expansion for

breakpoint n given by the following tk(0) + α[tk(n)− tk(0)] where α is the expansion factor.

In “proportional“ expansion/contraction mode, the new time is relative to two fixed time

boundaries Tmin and Tmax. Breakpoints that fall before Tmin are left unchanged. Breakpoints

falling between Tmin and Tmax are scaled. Breakpoints the occur after Tmax are shifted

earlier (for contraction) or later (for expansion). The following equation summarizes:

t′k(n) =

tk(n) for tk(n) ≤ Tmin

Tmin + α[tk(n)− Tmin] for tk(n) > Tmin ∧ tk(n) ≤ Tmax

tk(n) + (α− 1) [Tmax − Tmin] for tk(n) > Tmax

Proportional expansion/contraction scales the time base uniformly and accomplishes the

same thing as traditional time warping. Independent expansion/contraction can produce

interesting effects that “smear“ and overlap the evolution of each individual partial (in the

case of expansion), or emphasize short sinusoidal segments (in the case of contraction).

Figures 4.6–4.8 on page 71 illustrate the different time expansion modes.

Page 81: Klingbeil Dissertation Web

4. Editing and Transformation 71

Figure 4.6: Original partials prior to time expansion.

Figure 4.7: Partials time expanded in independent mode. Note that the start time of eachpartial is the same as in the unexpanded version.

Figure 4.8: Partials time expanded in proportional mode. Note that the start time of eachpartial is scaled relative to time zero.

Page 82: Klingbeil Dissertation Web

4. Editing and Transformation 72

4.5 Frequency Modifications

Currently the SPEAR GUI only supports a few basic frequency-based transformations.

Frequency shifting and transposition can be performed either by dragging the selection

vertically with the appropriate tool or by entering a specific frequency shift (in hertz) or

transposition (in floating point semitones). The interface is shown in figure 4.9

Figure 4.9: Dialog boxes for transposition and frequency shifting.

Frequency “flipping“ inverts all frequencies in the range Fmin to Fmax around the

axis 12 (Fmin + Fmax). For frequency fk(n) of partial k and breakpoint n, the new flipped

frequency f ′k(n) is given by

f ′k(n) = Fmax − ( fk(n)− Fmin)

Frequency flipping can be effective when applied to specific frequency bands. For example,

one might wish to try flipping the 3rd through 11th partials of an inharmonic timbre.

Figure 4.10 illustrates the effect. Frequency flipping tends to be less interesting when

applied to an entire instrumental sound, as the strong lower partials invert into high

frequency regions (often above 10 kHz). Strong energy in this area is often perceived

simply as a loud noise band.

4.6 Data Exchange

A central feature of SPEAR is support for data exchange with existing analysis synthesis

packages. For example, one might wish to use SPEAR’s visualization to compare the

analysis results from different software. The Sound Description Interchange Format (SDIF)

(Wright et al. 1999a) is supported both for import and export.

Page 83: Klingbeil Dissertation Web

4. Editing and Transformation 73

(a) Selected partials of a bell sound prior to frequency flipping.

(b) Selected partials of a bell sound after frequency flipping.

Figure 4.10: Frequency flipping.

SDIF was developed out of a need for a standard file format that could represent

sounds in ways other than as a sequence of time domain samples. SDIF is particularly

well-suited to storing spectral models. An SDIF file consists of a sequence of frames that

are tagged with a 64-bit floating point time value (measured in seconds). Each frame is

also tagged with a 32-bit integer stream ID which allows multiple logical channels per

file. Each frame consists of one or more matrices that contain the data. Various standard

matrix types, which are tagged with a 4-byte code, have been defined for different types of

analysis data and are shown in table 4.1.

For SDIF import, SPEAR supports 1TRC (sinusoidal tracks), 1HRM (harmonic sinu-

soidal tracks), RBEP (reassigned bandwidth enhanced partials — a matrix type used by

Page 84: Klingbeil Dissertation Web

4. Editing and Transformation 74

Matrix Type Description1FQ0 Fundamental frequency estimates1STF Short time Fourier transform data. The time tag

in a 1STF frame is the time of the center of thewindow, not the beginning.

1PIC Picked spectral peaks with amplitude, frequency,phase, and confidence values

1TRC Sinusoidal tracks with track index, amplitude,frequency, and phase values

1HRM Sinusoidal tracks with track harmonic number,amplitude, frequency, and phase values

1RES Exponentially decaying sinusoids/resonanceswith amplitude, frequency, decay rate, and ini-tial phase values

1TDS Time domain samplesRBEP Reassigned, bandwidth enhanced sinusoidal

tracks with track index, amplitude, frequency,phase, time-offset, and noise values

1ENV Sampled spectral envelope1NVT Name-value table storing text or numerical

header information (value) for arbitrary keys(name)

EASM Matrix storing all breakpoints of single partial.Supported by SDIF-Edit (Bresson and Agon 2004).

Table 4.1: Some of the standard SDIF matrix types.

Loris), and 1STF (short-time Fourier transform frames). Data can be exported either as

1TRC frames or RBEP frames. In the case of 1TRC frames, resampling of the breakpoint

functions must be performed to conform to the fixed frames of the 1TRC file type. For

RBEP frames, each point has a time offset value. By collating all breakpoints into frame-

sized chunks and properly setting the time offset, any distribution of breakpoints can be

losslessly represented with an RBEP SDIF file.

Additional formats are supported for importing — the .mq and .an formats from

SNDAN (Beauchamp 1993) and the .ats format from the ATS package (Pampin 2004). The

data exchange options are summarized in table 4.2.

Two custom text file formats have been implemented for both import and export.

par-text-frame-format represents resampled frames, where each line of the file con-

Page 85: Klingbeil Dissertation Web

4. Editing and Transformation 75

Format Import ExportSDIF 1TRC × ×SDIF RBEP × ×SDIF 1HRM ×SDIF 1STF ×SNDAN .mq ×SNDAN .pv ×ATS .ats ×Text frames × ×Text partials × ×

Table 4.2: Supported file formats.

tains the data of a single time frame. Index numbers are used to link sinusoidal tracks.

par-text-partials-format represents exact breakpoint functions, where each line of the

file represents a complete partial. Each line is separated by a newline character. Both for-

mats begin with a preamble. The first line of the preamble is either par-text-frame-format

or par-text-partials-format depending on the type of file. The second line specifies the

data types for each breakpoint. Subsequent lines give data about the number of partials

and/or frames. The final line of the preamble is either frame-data or partials-data de-

pending on the type. Figures 4.11–4.12 show the frame based format and figures 4.13–4.14

show the partials based format.

Although the SPEAR graphical interface is intuitive and fairly feature complete, it

may not offer enough flexibility for more advanced uses. For example, batch analysis and

transformation of many sound files would be quite tedious. The ability to import and

export data gives motivated users the option to implement batch transformations with

other applications (see section 5.3 for details). The motivation for creating a GUI based

analysis/synthesis tool arose from the dearth of options. Existing options such as SNDAN,

ATS, Loris, and IRCAM’s Additive were all primarily command-line based applications.

The ideal environment would offer strong graphical and scripting capabilities, and it is

hoped that future versions of SPEAR will incorporate a scripting engine to facilitate a

wider variety of use cases.

Page 86: Klingbeil Dissertation Web

4. Editing and Transformation 76

par-text-frame-format

point-typeindexfrequencyamplitude

partials-count<J>

frame-count<K>

frame-data

<frame-time0><N><index0><freq0><amp0>...<indexN><freqN><ampN>

<frame-time1><N><index0><freq0><amp0>...<indexN><freqN><ampN>

...

<frame-timeK><N><index0><freq0><amp0>...<indexN><freqN><ampN>

Figu

re4.

11: T

hepar-text-frame-format

spec

ifica

tion

.Fol

low

ing

theframe-data

line,

each

line

cont

ains

the

brea

kpoi

nts

for

one

fram

e.N

indi

cate

sth

enu

mbe

rof

peak

sin

each

fram

e.Th

eindex

valu

esco

nnec

tpea

ksfr

omfr

ame

tofr

ame.

Each

line

isse

para

ted

bya

new

line

char

acte

r.

par-text-frame-format

point-typeindexfrequencyamplitude

partials-count5

frame-count7

frame-data

0.0000005479.2095490.0002713199.7149510.0005512350.3475040.0025471540.6500850.0002620708.7898560.000477

0.0100005477.8322070.0004533199.7565460.0009492351.6083070.0139541529.9029540.0006250704.9032590.004945

0.0200005475.4076000.0005253201.2115630.0011512352.6972960.0345721533.9382930.0008910701.3994140.020578

0.0300005474.9185870.0004093199.9461820.0011822352.3843990.0514921552.4293210.0010800702.3543090.042789

0.0400005475.1724780.0001743198.6365660.0011092352.0135800.0596331564.0255740.0006420704.8601680.063039

0.05000033199.0298310.0010012353.0664370.0612330706.2393190.076010

0.06000033199.6420290.0009072355.6127930.0613630706.9907230.080715

Figu

re4.

12:S

ampl

eda

taas

par-text-frame-format

.The

rear

ea

tota

lof

5pa

rtia

lsan

d7

fram

es.

Page 87: Klingbeil Dissertation Web

4. Editing and Transformation 77

par-text-partials-format

point-typetimefrequencyamplitude

partials-count<J>

partials-data

<index0><P><start-time><end-time>

<time0><freq0><amp0>...<timeP><freqP><ampP>

<index1><P><start-time><end-time>

<time0><freq0><amp0>...<timeP><freqP><ampP>

...

<indexJ><P><start-time><end-time>

<time0><freq0><amp0>...<timeP><freqP><ampP>

Figu

re4.

13:T

hepar-text-partials-format

spec

ifica

tion

.Fo

llow

ing

thepartials-data

line,

each

pair

oflin

esco

ntai

nsth

ed

ata

for

one

part

ial.

The

first

line

ofea

chpa

irgi

ves

over

alld

ata

for

each

part

ial:

inde

xnu

mbe

r,le

ngth

(P),

star

ttim

e,an

den

dtim

e.Th

ese

cond

line

give

sth

ebr

eakp

oint

s.Ea

chlin

eis

sepa

rate

dby

ane

wlin

ech

arac

ter.

par-text-partials-format

point-typetimefrequencyamplitude

partials-count5

partials-data

050.0000000.046440

0.000000540.6500850.0002620.011610528.1726680.0006830.023220536.1510620.0009700.034830564.0255740.001158...

170.0000000.069660

0.000000350.3475040.0025470.011610351.8112790.0157910.023220353.0373230.0417800.034830351.9192810.058412...

270.0000000.069660

0.000000199.7149510.0005510.011610199.7632450.0010130.023220201.7673950.0012040.034830198.6487880.001166...

350.0000000.046440

0.00000079.2095490.0002710.01161077.6104580.0004830.02322074.5621800.0005420.03483075.1724780.000314...

460.0116100.069660

0.011610704.2775270.0056640.023220700.2948000.0263020.034830703.8214720.0545330.046440706.1539310.073634...

Figu

re4.

14:S

ampl

eda

taas

par-text-partials-format

.The

rear

ea

tota

lof

5pa

rtia

ls.

Page 88: Klingbeil Dissertation Web

78

5. Compositional Applications

The introductory chapter outlined a number of challenges composers face when working

with the computer: the time/cost problem, the synthesis problem, the control problem,

and the compositional problem. Thus far we have seen that SPEAR most directly addresses

the time/cost problem (by providing fast synthesis) and the control problem (by providing

an agile and highly intuitive user interface). We now turn in earnest to the composition

problem. Musical tools, such as SPEAR, are never neutral agents in the compositional

process. Whether deliberate or not, musical intentions are always latent in the design of

computer music software systems. SPEAR was conceived with particular attention to the

needs of spectral composition.

5.1 Spectral Composition

Spectral composition might be seen as subset of the more general notion of timbral

composition. Many of the notable practitioners of spectral composition are quick to eschew

the idea of a specific “spectral style.” Nevertheless, developments over the past thirty

years do point to consistencies which allow us to formulate a coherent conception of

spectral music. The wariness of labels stems in part from a mistrust of compositional

orthodoxies, serialism in particular, from which many spectral composers were trying

to escape. A central tenet of spectral composition in fact centers around the rejection of

discrete categories in favor of continuums. This distinguishes the spectral approach from

other notions of timbral composition which are more oriented toward categorical and

hierarchical notions: Fred Lerdahl’s Timbral Hierarchies (1987), Pierre Schaefer’s Solfège de

l’objet sonore, or James Dashow’s Dyad System (1999).

Page 89: Klingbeil Dissertation Web

5. Compositional Applications 79

Spectral composition can be understood from several perspectives. From an aesthetic

point of view, the notion of continuity and process is central. Tristan Murail speaks in

particular about the organization of continuous frequency space:

“Frequency space is continuous and acoustical reality only has to define itsown temperaments. If we push this reasoning to an extreme, the combinationof pure frequencies could be used to explain all past categories of musicaldiscourse and all future ones. Harmony, melody, counterpoint, orchestration,etc., become outdated and are included in larger concepts.” (Murail 1984)

Gérard Grisey holds a complementary view that locates notions of continuity and process

in the temporal domain.

“For me, spectral music has a temporal origin. It was necessary at a particularmoment in our history to give form to the exploration of an extremely dilatedtime and to allow the finest degree of control for the transition from one soundto the next. . . . From its beginnings, this music has been characterized bythe hypnotic power of slowness and by a virtual obsession with continuity,thresholds, transience and dynamic forms.” (Grisey 2000)

Joanthan Harvey’s view of spectralism stresses the potentially spiritual aspect of the

approach: “History seems grand, for once; spectralism is a moment of fundamental shift

after which thinking about music can never be quite the same again. . . . spectralism

in its simplest form as color-thinking, is a spiritual breakthrough.” Later he comments

specifically:

“In several works, to take a simple example, violins provide upper harmonicsto a louder, lower fundamental and at a given moment they cease to fuse, beginto vibrate, begin to move with independent intervals and then again return totheir previous state. The images of union and individuation are powerful oneswhich have both psychological and mystical implications. ‘The Many and theOne.’” (Harvey 1999)

It is notable that all three composers make a particular point of spectralism’s place

in history. Composers of the postwar generation seemed particularly keen to define

spectralism as a coherent alternative to what they perceived as the cognitively opaque

results of serialism. Spectral composition nevertheless gravitates toward a formalist

approach that privileges individual compositional control. Technology has tended to

Page 90: Klingbeil Dissertation Web

5. Compositional Applications 80

function as a means of gaining precision and power. It remains an interesting open area

for exploration to see how spectral approaches might be applied in improvisational or

indeterminate compositional situations.

From a technical point of view, spectral music engages a number of specific concerns.

Pitch is organized in a continuous fashion and pitch structures are often reckoned directly

in cycles per second. In instrumental writing, approximations to the nearest quarter-

tone (or in some cases eighth tone) are used. This is in marked contrast to a precise

just intonation pitch specification.1 The harmonic series serves as an important point of

departure, and often functions as a stable pole (if not the stable pole) in the harmonic

discourse. Harmony often proceeds by a differential approach which unfolds processually.

The notion of “instrumental additive synthesis” often features prominently in spectral

composition. A combination of instrumental sounds with the appropriate frequencies and

intensities can fuse into a single timbral entity. In contrast to some approaches to timbral

composition, instrumental color tends to be exploited not for its differentiating abilities

but for its potential to participate in spectral fusion. Thus individual instrumental timbres

are subsumed into larger musical structures.2 In some spectral works, harmonic models

are derived from characteristic spectra of instrumental sounds. The goal of instrumental

additive synthesis is never to recreate, but rather to reveal latent musical potential is

pre-exisiting sonic material. The archetypal gesture of instrumental additive synthesis is

the arpeggiated spectra which gradually transforms from a series of individual tones to a

fused complex.

As with pitch, the approach to rhythm derives from a temporal sensibility that is

continuous rather than discrete. Rhythmic processes often involve controlled long range

accelerations and decelerations. Additive and subtractive rhythms or rhythmic interpola-

tions (as in the music of Philippe Hurel) are another common strategy. Overlaid rhythmic

1The approximation approach tends to be found in the music of the French spectralists and their followers.American music that might be identified as spectral or proto-spectral has tended to follow a just intonationapproach (as in the compositions of James Tenney or Ben Johnston).

2This technique is a hallmark of spectral music. While instrumental timbre played an important role insome proto-spectral music (that of Varèse in particular) timbre tended to be more about making elementsmore distinct (Varèse’s planes of sound) rather than fusing them together.

Page 91: Klingbeil Dissertation Web

5. Compositional Applications 81

pulsations with various rational relationships are also encountered (particularly in the

music of Grisey). In many cases rhythms are approximated, either via quantization or

proportional notation.

A complete summary of all technical aspects found in spectral music is beyond the

scope of this document (see Fineberg (2000) for more details). For our discussion, the

following list of some of the central concerns of spectral music will suffice:

• continuous frequency space

• continuous temporal space

• timbre-harmony continuum

• the use of approximations

• global approach to form

• smooth transitions contrasted with ruptures

• process preferred over development

• appeal to perceptual and psychoacoustic aspects of sound

The last item points to a particularly anti-formalist aspect of spectral music. For all

its technical machinations, spectral music is very much “for the listener.” However, this

tendency to appeal to cognitive “universals” raises interesting concerns. For example,

to what degree might the expressive or communicative ability of the spectral approach

be limited by this focus? In particular, does the pre-occupation with the purely acoustic

preclude the possibility of addressing culture-specific aspects of the musical enterprise?

These open questions point to interesting possible directions for future compositional

work.

Technology has always played an important role in the development of spectral music.

Models derived from electronic processes such as frequency shifting, ring modulation,

frequency modulation, or tape delay can be found in any number of spectral pieces.

The desire for mastery of the sonic continuum led naturally to the use of the digital

computer.3 Digital sound synthesis is complementary to instrumental synthesis and

3Ironically, one gains reproducible control of the continuum in the discrete world of the computer.

Page 92: Klingbeil Dissertation Web

5. Compositional Applications 82

spectral works that combine electronic and instrumental sounds can achieve remarkable

coherence. Computer-assisted composition environments, which will be discussed in

section 5.3, are indispensable for realizing sonic processes. SPEAR is a particularly useful

tool for spectral composition. The high performance GUI makes it possible to visualize

and manipulate complex sounds. Its detailed analyses can be used in the derivation of

harmonic and/or rhythmic models, or the data can be harnessed directly to drive various

forms of digital synthesis.

5.2 Sonic Transformations

A sinusoidal analysis may be viewed as a reservoir of sonic data with almost limitless

musical potential. Possible temporal transformations include replication, time delay, time

dilatation, and time reversal. Amplitude envelopes may be applied (for example, to fade

partials in and out, or to emphasize the attack portion of a sound). Partial frequencies may

be altered by any desired transposition or frequency shifting envelope. The analysis data

can be used to drive other compositional or synthesis processes (for example, granular

synthesis). The following subsections detail a number of possible sonic transformations.

5.2.1 Timbre Hybridization

Properties of different spectral analyses can be combined to create new hybrids. This

process of timbre hybridization is sometimes referred to as cross synthesis. There are many

possible cross synthesis implementations, some of which are found in software such as

AudioSculpt, SoundHack, Shapee, and in commercial plugins such as Prosoniq Morph.

Many of these techniques focus on different ways of combining STFT frames or applying

properties of one STFT frame to another. Sinusoidal modeling offers the possibility of very

fine-grained control over the cross synthesis process.

Dynamic morphing between two collections of partials A and B can be achieved by

matching common partials and interpolating their frequencies and amplitudes. Such a

procedure has been implemented as in Common LISP as follows:

Page 93: Klingbeil Dissertation Web

5. Compositional Applications 83

1. Temporally align the two sounds by means of timebase envelopes that map output

times to time points in the original sound.

2. Define an interpolation envelope that maps output time to relative fraction of A or

B that should be present in the output — a value of 0 indicates the sound should

consist of all A, 1 indicates all B, intermediate values should be a blend of A and B.

3. Compute a succession of blended spectral frames to be synthesized. This proceeds

as follows:

• Advance an output time variable t by a predetermined hop.

• For each value of t, determine the active partials in A and B and compute two

frames At and Bt with the interpolated frequencies and amplitudes of the active

partials at time t.

• For each value of t, determine the current blend factor x from the interpolation

envelope.

• Determine a set of matching frequency/amplitude points in At and Bt. Matching

is controlled by a frequency tolerance. Interpolate the amplitude and frequency

of each of the matching points according to the blend factor x. Unmatched

points can be discarded or have their amplitude adjusted according to x (thereby

creating a cross fade of the unmatched partials).

4. Synthesize the blended frame and window with a tapered window function. Overlap

add into the output.

This morphing algorithm is somewhat crude, but is reasonably effective for certain kinds

of sounds. It could be improved by computing a new set of blended partials, rather than

overlap adding phase-incoherent frames in a granular fashion. As it is, the algorithm only

works well if sounds contain similar frequency content. Harmonic sounds need to share

the same fundamental frequency. A more sophisticated algorithm, such as that used in

Loris, would tag the harmonics of A and B relative to a fundamental frequency track.

Like harmonics can then be interpolated and smoothly glide to new frequencies. For the

specifics of the Loris morphing algorithms see Fitz et al. (2003).

Page 94: Klingbeil Dissertation Web

5. Compositional Applications 84

A more sophisticated algorithm should also be able to interpolate between more than

two sources. It would be musically useful to create long sequences of seamless morphs

between a series of sounds as with IRCAM’s Diphone Studio. Despite its limitations, this

basic algorithm shows promise and has been used to create some effective morphs between

harmonic and inharmonic timbres.

Another approach to cross synthesis is to keep the frequency content fixed and ma-

nipulate amplitudes to affect a change in the spectral envelope (see section 2.3 for a

brief discussion of spectral envelopes). A hybrid can be created by imposing the spectral

envelope of one sound on another. Although SPEAR does not currently include spectral

envelope capabilities, a number of Common LISP functions have been developed for

manipulating spectral envelopes and applying them to SPEAR analyses (see table 5.3). The

functions are designed to work with spectral envelope data generated by SuperVP, a pow-

erful command line analysis/synthesis tool developed at IRCAM. Successful experiments

have been conducted with the application of vocal spectral envelopes to tam-tam sounds

(and visa-versa).

5.2.2 Spectral Tuning

Spectral tuning adjusts the frequency content of a sound to match the frequencies of a

predetermined spectrum or harmonic field. The target frequencies might be derived from

any number of sources: octave repeating scales or chords, synthetic spectra, frequencies

derived from other spectral analyses, distortion and warping of existing spectra, etc.

Spectral tuning using the phase vocoder has been explored extensively by a number of

composers and researchers including Eric Lyon and Christopher Penrose with FFTease

(Lyon 2004), Paul Koonce with PVC, and Trevor Wishart with Composers Desktop Project

(CDP) (Wishart 2000).

Spectral tuning with a sinusoidal model allows very precise control of the evolution of

each partial. In a typical application, the tuning is static: the local amplitude and pitch

fluctuations of each partial are maintained and the partial is transposed so that its average

Page 95: Klingbeil Dissertation Web

5. Compositional Applications 85

frequency matches the target. Local pitch fluctuations can be attenuated or emphasized

as desired. Dynamic tuning transposes a partial according to a variable function. If the

rate of frequency change is not too rapid, the retuning can still maintain local frequency

perturbations which are crucial to the character of most natural sounds. A simple example

of dynamic tuning would be the application of vibrato or glissando.

Parameters can control the strength of retuning. For example if a partial is separated by

a distance of 3 semitones from its target, a retuning strength of 70% would transpose the

target by 2.153 semitones rather than the full 3 semitones. Or a parameter can specify that

partials are retuned only if they are in close enough proximity to the target. Particularly

interesting results have been achieved with progressively retuning inharmonic timbres so

that a succession of sounds eventually converges on a harmonic spectrum. Spectral tuning

can also be used for cross synthesis where the target frequencies are derived from analysis

of a different sound.

5.3 Software Composition Environments

Although a great deal of useful sonic manipulation and transformation can be accom-

plished directly with the SPEAR GUI, much of the software’s compositional utility is

revealed when it is used in conjunction with other tools. Computer-assisted composition

and performance environments such as OpenMusic, Common Music, Max/MSP, RTCmix,

and SuperCollider (to name a few), can be used to import, manipulate, export, and/or

synthesize SPEAR analysis data.

SPEAR’s support for text based file formats (as well as standard SDIF formats) makes

cross program data exchange relatively straightforward. The text formats are particularly

useful because they are human readable and easy to use with interactive interpreted

programming languages such as Java, Python, Scheme, and Common LISP. Import and

export of the SDIF formats is more difficult and time consuming to implement for casual

programmers. The author’s own experience has shown that even though SDIF is robust and

well-defined, it is not necessarily widely understood among computer music practitioners.

Page 96: Klingbeil Dissertation Web

5. Compositional Applications 86

Even with supporting SDIF libraries such as those provided by CNMAT or IRCAM, it may

prove more expedient (particularly in one-off applications) to use the text data formats.

5.3.1 OpenMusic

OpenMusic is a computer-assisted composition environment developed at IRCAM using

the Common LISP language. It is a visual programming environment based on patching of

pre-defined or user-defined objects. The patch paradigm provides full access to Common

LISP programming, including iteration and recursion. Nodes in the object graph are

evaluated for their results and possible side-effects. A user request for evaluation will

trigger evaluation of all parent nodes (figure 5.1). OpenMusic includes graphical editors for

MIDI data, breakpoint functions, and common practice notation (figure 5.2). OpenMusic

was developed as a successor to PatchWork a similar LISP based environment for visual

programming.

Figure 5.1: OpenMusic patch that evaluates the arithmetic expression 10 + (2× 10) .

The environment may be extended with user defined libraries. New OpenMusic func-

tion are defined as CLOS (Common LISP Object System) methods using the om:defmethod!

macro. SPEAR import and export has been integrated with OMTristan, a custom OM

library created by Tristan Murail. OMTristan, which has evolved over decades, includes a

wide variety of compositional algorithms for the generation of chords and spectra, tempo-

ral computations, MIDI communication, and general data manipulation. The spdata object

in OMTristan is specifically designed for efficient storage and manipulation of frame based

spectral data. The methods spear-read and spear-write convert between the SPEAR

par-text-frame-format file type and OM spdata objects.

Page 97: Klingbeil Dissertation Web

5. Compositional Applications 87

Figure 5.2: OpenMusic notation editor displaying a chord sequence.

Figure 5.3 shows an OM patch that reads SPEAR data, transforms the data, displays

the transformed data in common practice notation, and writes out the transformed data

to a new text file. The patch functions as follows: The spear-read object (located at the

top) takes the name of a text file as input and outputs an spdata object which is stored

as an embedded object (this avoids having to re-read the data from disk on subsequent

evaluations). The spdata output is split into three streams: the frequencies, the amplitudes,

and the partial index numbers. Each stream is a LISP list of two elements. The first element

is a list of times for each frame and the second element is a list of lists where each sublist is

the frame data (either frequencies, amplitudes, or indices). The time base is stretched (the

“dilation” parameter) and the frequencies distorted in the dist-frq sub-patch. The data is

reassembled into frames and converted to a chord sequence in the visu-chseq sub-patch.

The times, indices, frequencies, and amplitudes are also routed to the spear-write object

which outputs a new text file with the transformed data. The data can then be opened in

SPEAR for resynthesis or further transformation.

5.3.2 Max/MSP

Max/MSP is a widely used graphical patching environment for realtime media processing.

SPEAR SDIF data can be imported into Max/MSP using the CNMAT SDIF objects (Wright

et al. 1999b). The SDIF-buffer stores the SDIF data in a named buffer which can be

accessed using the SDIF-tuples object. SDIF-tuples outputs matrix data at the specified

Page 98: Klingbeil Dissertation Web

5. Compositional Applications 88

Figure 5.3: OpenMusic patch that reads and writes SPEAR data.

Page 99: Klingbeil Dissertation Web

5. Compositional Applications 89

time (with optional interpolation) as a Max list. CNMAT’s sinusoids~ object implements

an oscillator bank (with optional bandwidth enhancement as described in section 3.6.1)

that is driven by a list of frequency amplitude pairs. The threefates object can be used to

manage the sinusoidal track indices output from the SDIF buffer and format the data into

frequency amplitude pairs suitable for the sinusoids~ object.

shakuhachi

•!start audio

startwindow stop

SDIF-buffer ss-toot

SDIF-tuples ss-toot

sinusoids~

tuples time $1 columns 1 2 3

read-stream-number moanin.add.sdif 0

loadbang

dac~

SDIF-tuples ss-boot

tuples time $1 columns 1 2 3

2.1

line 0. 10.

0., 10. 10000 7.1, 2.1 5000

sdif-buffer ss-boot

read-stream-number kyorei.add.sdif 0

loadbang

list-interpolate

zl slice 100zl slice 100

t b l

f 1.

0.

t f f

threefates threefates

sax

While this fast and dirty interpolation yields some

results, it would be more musical if we paid

attention to the actual meaning of the SDIF data.

List-interpolate with sines

sin-transform

prepend sinusoids frequency-scale $1

0.

Trying to interpolate between timbre, contour, and

amplitude envelope all it once is a crazy idea, if

you think about it.

Figure 5.4: Max/MSP patch from the CNMAT Spectral Synthesis Tutorials that interpolatesbetween two sinusoidal models to achieve a dynamically controllable cross synthesis.

The limitation of 256 elements with the standard Max list objects does put some

constraints on the complexity of the spectral data. Michael Zbyszynski’s “Spectral Synthesis

Tutorials” suggest some solutions (Zbyszynski et al. 2007). Possibilities include the use of

custom C or Java objects, the use of Javascript, the use of Jitter matrix processing objects,

or the use of IRCAM’s FTM extensions.

Page 100: Klingbeil Dissertation Web

5. Compositional Applications 90

The use of sinusoidal models in the Max/MSP/Jitter environment offers numerous

options for interactive transformation and resynthesis. The ability to manipulate spectral

data directly in Max using Java and Javascript eases the burden of managing large spectral

data sets. Future work could lead to the development of “partial based” storage models

rather than the frame based approach implicit in the SDIF 1TRC format. As we shall

see in the next section, the partials based format (collections of breakpoint functions) is

particularly well suited for certain types of musical operations.

5.3.3 Common Music

Common Music (CM) is an algorithmic composition environment developed by Heinrich

Taube. Like OpenMusic, CM is implemented in Common LISP. It is built around the notion

of generalized input/output streams which can include MIDI files, MIDI ports, Open

Sound Control streams, CSound scores, CLM notelists, and CMN scores. Compositional

facilities include the scheduled processes, a full featured pattern iteration system, and a

flexible graphical interface called Plotter. Plotter facilitates display and editing of breakpoint

functions, histograms, scatter plots, and piano roll style MIDI data (Taube 2005). CM can

be tightly integrated with CLM (Common Lisp Music), a Music-N style acoustic compiler.

Current versions of CM include a spectral composition toolkit than can read SPEAR

data from par-text-frame-format and par-text-partials-format files. The data is stored

as nested LISP lists. For manipulating large sounds, it can be more efficient to store spectral

data in LISP vectors (contiguous arrays). A LISP toolkit has been developed for reading,

writing, and manipulating par-text-partials-format data. A complete sound is stored as

a LISP list of sp-partial data structures. Each sp-partial represents a single breakpoint

envelope with time, frequency, and amplitude arrays. Functions exist to transpose, time

warp, and amplitude scale partials. The adjustments can be controlled dynamically with

envelopes. The function for transforming single partials are summarized in table 5.1, page

91. Many of the transformation functions take an optional copy parameter which, when

Page 101: Klingbeil Dissertation Web

5. Compositional Applications 91

(copy-partial partial)

Return a newly allocated copy of partial.

(offset-partial partial offset [copy])

Shift the time of partial by offset seconds.

(transpose-partial partial interval-ratio [copy])

Transpose the frequencies of partial by interval-ratio.

(transpose-partial-env partial interval-ratio strength-env [copy])

Transpose the frequencies of partial by a ratio ranging from 1 to interval-ratio. strength-envis an envelope controlling the percentage of interval-ratio applied.

(time-scale-partial partial timescale [copy])

Apply a fixed time scaling ratio timescale to partial.

(amplitude-scale-partial partial ampscale [copy])

Apply a fixed amplitude scaling ratio ampscale to partial.

(reverse-partial partial [copy])

Reverse the temporal evolution of partial.

(amp-env-partial partial env [copy])

Apply an amplitude scaling envelope env to partial.

(sp-partial-avg-freq partial)

Compute the average frequency of partial.

(sp-partial-avg-amp partial [points])

Compute the average amplitude of partial. If points is non-nil, the average is computed based onthe first points breakpoints. If points is nil (the default) all breakpoints are used to compute theaverage.

(sp-partial-interp partial time)

Interpolates the frequency and amplitude of partial at time seconds. The frequency and amplitudeare returned as multiple values (in that order). Uses a binary search for efficiency.

Table 5.1: Transformation functions applied to individual partials.

non-nil, applies the transformation to a new copy of partial, leaving the original unchanged.

The default value of copy is nil, meaning the transformation is applied destructively.

A retune-partials function can transpose a collection of partials to a new set of

target frequencies. The retune-partials function assumes partials with relatively constant

frequency (the retuning is based on the average frequency of the original partial). Functions

Page 102: Klingbeil Dissertation Web

5. Compositional Applications 92

(retune-partials partials target {keyword value}*)

Transpose each of the partials of partials so that its average frequency matches a frequencycontained in the list target. Because the target frequency set may cover a vastly different frequencyrange from the original partials, the average partial frequencies are first rescaled to lie within theminimum and maximum frequencies of target. Each rescaled frequency is then matched to theclosest frequency in target. The matching is controlled by the following keyword arguments::low-limit Only tune partials above this frequncy. Default is 0 Hz.:high-limit Only tune partials below this frequcny. Default is 3000 Hz.:delay-curve Envelope to apply a frequency dependent delay to each partial. De-

fault is nil.:amp-curve Envelope to apply frequency dependent amplitude scaling to each

partial. Default is nil.:flatten Scale the ambitus of each partial. Values < 1.0 reduce the amount of

frequency variation of each partial. When 0.0, all frequencies are setto the average frequency.

:strength Adjust the amount of frequency remapping. When 1.0, the partialsare retuned to exactly match the target frequencies. When 0.0, thepartials remain at their original frequencies.

:strength-curve Adjust the amount of frequency remapping over time. Partials can bemade to glide between original and target frequencies.

:index-scramble Scramble the targets such that the original partials will remap toless proximate target frequencies. When 0, there is no scrambling(the default). When 1, remap to the neighbor of the closest targetfrequency. When n, remap to the nth neighbor (either above orbelow).

(sp-partials-apply-vibrato partials {keyword value}*)

Apply slow frequency modulation (vibrato) to partials (a list of partials). This function requiresCLM. The following keyword arguments are supported::vibfreq The frequency of the vibrato in hertz. Default is 6.5.:vibinterval The width of the vibrato in semitones. Default is 0.5.:start-time Starting time for the onset of vibrato (in seconds). Default is 0.:end-time Ending time for the vibrato (in seconds) or nil if vibrato continues to

the end. Default is nil.:copy If non-nil, return new copies of the partials leaving the originals

unchanged. Default is true.

Table 5.2: Frequency adjustment functions applied to collections of partials.

for frequency transformation are summarized in table 5.2, page 92. Spectral envelope

functions are show in table 5.3, page 93.

5.4 Compositional Examples

A straightforward application of SPEAR is the derivation of harmonic models for instru-

mental composition. Pitches can be transcribed manually from the GUI or by importing

Page 103: Klingbeil Dissertation Web

5. Compositional Applications 93

(sp-partials-flatten-spectral-env partials spectral-env [amp-scaler])

Flatten the spectral envelope of the sound represented by partials by dividing the partial ampli-tudes by the spectral envelope spectral-env. The spectral envelope can be generated by readingan SVP text formatted spectral envelope analysis. Apply an optional amplitude scaler amp-scaler(default value 0.025) following the division.

(sp-partials-apply-spectral-env partials old-spectral-env new-spectral-env {keywordvalue}*)

Apply a new spectral envelope new-spectral-env to the sound represented by partials.old-spectral-envelope is the current spectral envelope of the sound. The following keywordarguments are supported::transpose Transposition ratio.:time-warp Time remapping envelope.:freq-warp Frequency remapping envelope.:freq-func Generalized frequency transformation. freq-func is a function of

three arguments: frequency, partial-index, and time. The functionshould return the new desired frequency value.

:filter-func Generalized amplitude transformation. filter-func is a functionof four arguments: amplitude, frequency, partial-index, and time.The function should return the new desired amplitude value.

:copy If non-nil, return new copies of the partials leaving the originalsunchanged. Default is true.

Table 5.3: Spectral envelope functions applied to collections of partials.

into other environments. Rather than relying on individual spectral snapshots, the evolu-

tion of a collection of partials can be incorporated into the composition model.

In the author’s work Subterrain (2008) for B[ clarinet, string trio, and electronics,

clarinet multiphonics were analyzed. The SPEAR GUI was indispensable for extracting

only the most stable and/or musically interesting partials from these complex sounds.

The data was imported into Common Music and amplitudes were averaged across entire

partials, which gave a very good sense of the overall spectral color of the multiphonic.

Further manipulation of the multiphonic spectra, including frequency shifting and ring

modulation, were used to create harmonies and pitch reservoirs.

Spectral retuning using SPEAR and OpenMusic was used to create a number of sounds

in Tristan Murail’s Pour adoucir le cours du temps (2005) for 18 instruments and synthesized

sound. In this piece, the partials of tam-tam sounds were retuned to the frequencies of

artificially generated spectra. A cow-bell analysis also undergoes similar retuning. Of the

work, Murail writes,

Page 104: Klingbeil Dissertation Web

5. Compositional Applications 94

“The ‘noises’ (breath, grainy sounds, metallic resonances) of the piece havebeen domesticated, tuned to instrumental harmonies through the use of aspecific technique for altering the internal components of the sounds. This ishow the sound of a gong can become a harmony, or the virtual cow-bells canbe constantly varied and altered by their musical context. The synthesizedsounds are triggered from an onstage MIDI keyboard — they are mixed andspatialized in real-time by the program Holophon developed by the GMEM.”(Murail 2005)

Spectral tuning was used to generate several of the sounds used in the author’s work

Tear of the Clouds (2007) (see appendix A for the complete score). The low vocal sounds

(at m. 14, reh. A and m. 203) were created by retuning and resynthesizing samples of a

choral ensemble. Although these particular source samples exhibited a high standard of

performance and excellent recording quality, they were not ideally suited for resynthesis.

Because they consisted of a sum of many ensemble voices (all in unison pitch), and because

they were recorded in a reverberant environment, partial tracking resulted in a large

number of short low amplitude partials in addition to the fundamental and harmonics. For

retuning purposes, a cleaner representation was desired. IRCAM’s Additive program was

used to perform fundamental frequency tracking and to analyze a fixed set of harmonic

partials (up to 22 kHz). The SDIF 1TRC file from Additive was imported into SPEAR and

then exported as a text file. This was then imported into LISP where the retuning was

performed.

At m. 14, the retuning target consists of selected partials of a compressed harmonic

spectrum built on G1 (49 Hz). The tuning was only applied to the original partials ranging

from 49 Hz to 1850 Hz. Because the original sound had a fundamental of 82 Hz, retuning

and compressing the entire spectrum would have resulted in a considerable loss of high

frequency presence. Since most of the harmonic “color” of a sound is found below 2000

Hz, the harmonics above 1850 Hz were retained in their original tuning (another option

would be to replicate partials to fill in the missing upper harmonics). The retuned data was

then imported into SPEAR and synthesized. Further post processing, including granular

time stretching, was then applied. The granulation produced a sound of the required

Page 105: Klingbeil Dissertation Web

5. Compositional Applications 95

duration and helped re-impart some of the reverberant quality of the original sound that

was lost in the analysis.

Further applications of spectral tuning in Tear of the Clouds include the retuning of

tubular bell, piano, and tam-tam sounds. The penultimate electronic sound of the piece

is built from a retuning of a piano sound. The target spectrum consists of a distorted

harmonic series built on D]1 (39 Hz), in which the even partials are stretched and the

odd partials are unchanged. For this particular sound, the retuning is dynamic. The

attack maintains the original tuning and the partials glide to the target spectrum over a

period of 2.75 seconds. Rather than sliding to the nearest target frequency, the retuning

was configured so that the partials make wider frequency excursions (either up or down),

crossing each other as they proceed.

Figure 5.5: Dynamic retuning with crossing partials.

5.5 Performance Models

Although not extensively explored in the current work, some words are in order about the

potential applications in performance and improvisation. The SPEAR interface itself could

be viewed as a crude performance interface where one can scrub through the sound using

the mouse (Garton 2007). One can easily imagine a richer instrument that would give

Page 106: Klingbeil Dissertation Web

5. Compositional Applications 96

control of elements such as playback speed, transposition, frequency shifting, harmonic

parity (balance of even and odd partials), etc. These potentials would ideally be realized

in a flexible environment such as Max/MSP, rather than in the more confining space of

the SPEAR user interface. A variety of gestural and tactile inputs could be used to drive

resynthesis parameters.

Page 107: Klingbeil Dissertation Web

97

6. Conclusions

The preceding chapters have given a thorough account of the technical aspects of the

design and development of SPEAR while also touching on some possible compositional

applications. I do not wish to claim that SPEAR represents a revolutionary new computer

music tool. It has been preceded by more than fifteen years of research and development

in the field of spectral analysis and modeling. However, I hope it is clear that SPEAR is an

important evolutionary convergence of a number of mature technologies.

Although SPEAR has experienced reasonable popularity among a wide variety of

technologically oriented music practitioners, the software is still very much a work-in-

progress. Although the interface works well, it is still limited in many ways. A multi-

demensional display should be developed that allows one to view and edit all aspects

of a partial’s evolution: amplitude, noise bandwidth, phase, stereo (or multi-channel)

panning, etc. The user should be able to draw breakpoint envelopes and curves to control

the transformation of any parameter. Rule based selection (see section 4.2.1) should be

implemented. Moreover, an interface that allows one to specify persistent selections (like

audio regions in a waveform editor) would be incredibly useful. A user interface for

sound morphing and hybridization would also be welcome. Since it is impossible to

envision all of the possible desired editing and transformation possibilities, SPEAR needs

a built-in scripting language and command-line tools that allow users to develop their

own processes.

SPEAR must also keep pace with developments in spectral analysis and synthesis. To

build more compact and malleable spectral models, SPEAR needs robust fundamental

frequency tracking. The option should be available to perform analyses with fixed numbers

Page 108: Klingbeil Dissertation Web

6. Conclusions 98

of harmonics (as with IRCAM’s Additive and Diphone programs). Models with fixed

numbers of partials are much easier to manipulate in real-time environments such as

Max/MSP or SuperCollider.

Some form of noise modeling, as discussed in section 3.6, should be implemented. The

transient sharpening method described in section 3.5 is successful to a point, but to handle

wide-band polyphonic audio and percussive attacks, a separate transient detection and

representation method is needed. The multi-resolution sinusoidal modeling techniques

described in Levine (1998) should also be implemented. Although multi-resolution model-

ing significantly increases the implementation complexity of partial tracking (particularly

at the band boundaries), it has the potential to considerably reduce pre- and post-echo

artifacts. The combination of time-reassignment, multi-resolution analysis, noise modeling,

and improved partial tracking would represent a formidable analysis engine.

In order to tie together these more sophisticated analysis techniques, a more advanced

user interface is required. Each component of the analysis would be assigned to its own

layer:

• fundamental frequency layer

• sinusoids layer

• noise layer

• transient layer

• spectral envelope layer

Each of these layers could be (optionally) coupled to the others. For example, when

layers are coupled together, adjustments to the fundamental frequency track would

correspondingly adjust the frequencies in the sinusoids layer. As the partials are transposed,

the amplitudes could be adjusted to match the spectral envelope layer. Peak detection and

tracking could be applied to the spectral envelope data so that formant trajectories and

widths could be adjusted in a manner analogous to individual partials. Given sufficient

time and effort, all of these ideas are feasible.

Page 109: Klingbeil Dissertation Web

6. Conclusions 99

Returning to some of the issues raised in the opening chapter, we must consider what

a tool like SPEAR tells us about timbre and musical organization. For all its abilities,

SPEAR perhaps raises more questions than it answers. Although obvious, it’s worth

stating the overall conclusion: additive synthesis is not a sufficient model for timbre.

Although additive models are malleable, they tend to be fragile — minor adjustments to a

few partials can destroy timbral fusion and coherence. Higher level models are needed.

Spectral envelopes are a good first step. We should also look for models of the micro-level

evolution of partials. The amplitude and frequency fluctuations can be correlated between

partials and with parameters of the overall evolution of the sound. Finally, we must

recognize that the time varying aspects of spectra are often far more important than

the instantaneous relationships. We need timbral models than can capture micro-level

rhythm, grainy textures, and impulses. The TAPESTREA software, which uses a sinusoids

+ stochastic background + transients model, suggests a particularly promising approach to

high level timbre modeling (Misra et al. 2006).

From a compositional perspective there is clearly more to learn about timbre and

tools such as SPEAR can help with these compositional approaches. But it is not enough

for compositional work to merely reflect or reinforce habitual modes of understanding

timbre. Experimental composition can reveal new ways to listen to timbre, previously

unheard subtleties, surprising relationships, and unanticipated modes of organization.

With the acknowledgment that speculations on the future are often more about setting

an agenda than trying to make a neutral prediction, we might ask what role computer

music might play in this process and in turn what will be the nature of tools to come? Or

more precisely, what do we want from our musical tools of the future? Successful software

tends to embody at least some of the following attributes:

1. ease of use and configuration

2. high performance

3. large feature set

4. reliability

Page 110: Klingbeil Dissertation Web

6. Conclusions 100

5. backward compatibility

6. initial novelty

7. openness and extensibility

At present, commercial software tends to excel at items 1–3, while failing miserably in

terms of openness and user extensibility. Extensibility has typically been the strength

of software with an experimental and open source pedigree. Commercial software is

frustrating because it tends to be expensive and creatively limiting. Free software is

frustrating because it tends to be difficult to configure and maintain. Of what benefit is

extensibility if one is left with no time to exploit it? Future software efforts must bridge

this gap. We must avoid the “sealed black boxes” of the commercial world and the arcane

configuration, instability, and continual tabula rasa upheavals of the open source world.

Negative proscriptions aside, what potentials are there for music software of the

future? An environment that melds a powerful acoustic compiler and general purpose

programming language with a flexible and high-performance GUI seems ideal. Such an

environment should include flexible time-based displays that can handle common practice

notation, breakpoint functions, audio data, and spectral models. It must not be a neutral

environment — it must make assumptions. It must assume a user who wishes to work

with common practice diatonic music and quantized meter and rhythm; it must assume

a user who wishes to work with millions of sonic granules on a proportional time scale;

it must assume a user who wishes to precisely notate irrational rhythms and microtonal

inflections; it must assume a user who wishes to capture and manipulate performance data

in real time. Technologically, this is all within the realm of possibility today. It remains to

be seen when, if, and/or how this may come about. Tools of this sort are not envisioned to

make composition facile. They should exist to engage and inspire the human imagination

and creative spirit.

Page 111: Klingbeil Dissertation Web

101

References

Abe, Mototsugu, and Julius O. Smith. “Design Criteria for Simple Sinusoidal ParameterEstimation based on Quadratic Interpolation of FFT Magnitude Peaks.” In Proceedings ofthe 117th Audio Engineering Society Convention. AES, 2004.

Amatriain, Xavier, Maarten de Boer, Enrique Robledo, and David Garcia. “CLAM: AnOO Framework for Developing Audio and Music Applications.” In Proceedings of17th Annual ACM Conference on Object-Oriented Programming, Systems, Languages andApplications. Seattle, WA, 2002. http://mtg.upf.edu/publicacions.php.

Auger, F., and P. Flandrin. “Improving the Readability of Time-Frequency and Time-ScaleRepresentations by the Reassignment Method.” IEEE Transactions on Signal Processing 43:(1995) 1068–1089.

Beauchamp, James W. “Unix Workstation Software for Analysis, Graphics, Modification,and Synthesis of Musical Sounds.” In Proceedings of the 94th Audio Engineering SocietyConvention. AES, 1993. Preprint no. 3479.

Bonada, Jordi, and Xavier Serra. “Synthesis of the Singing Voice by Performance Samplingand Spectral Models.” IEEE Signal Processing Magazine 24, 2: (2007) 67–79.

Bresson, Jean, and Carlos Agon. “SDIF Sound Description Data Representation andManipulation in Computer Assisted Composition.” In Proceedings of the InternationalComputer Music Conference. Miami, FL: ICMC, 2004, 520–527.

Cadoz, Claude, Annie Luciani, and Jean Loup Florens. “CORDIS-ANIMA: A Modeling andSimulation System for Sound and Image Synthesis: The General Formalism.” ComputerMusic Journal 17, 1: (1993) 19–29.

Cormen, Thomas H., Charles E. Leiserson, and Ronald L. Rivest. Introduction to Algorithms.Cambridge, MA: MIT Press, 1990.

Dashow, James. “The Dyad System (Parts Two and Three).” Perspectives of New Music 37,2: (1999) 189–230. http://www.jamesdashow.net/download/DyadSystemPt2-3.pdf.

Depalle, Philippe, Guillermo García, and Xavier Rodet. “Tracking of Partials for AdditiveSound Synthesis Using Hidden Markov Models.” In IEEE International Conference onAcoustics, Speech and Signal Processing. Minneapolis, MN, 1993, 225–228.

Page 112: Klingbeil Dissertation Web

References 102

Fineberg, Joshua. “Guide to the basic concepts and techniques of spectral music.” Contem-porary Music Review 19, 2: (2000) 81–113.

Fitz, Kelly. The Reassigned Bandwidth-Enhanced Model of Additive Synthesis. Ph.D. thesis,University of Illinois Urbana-Champaign, Urbana, IL, 1999.

Fitz, Kelly, Lippold Haken, and Brian Holloway. “Lemur—A Tool for Timbre Manipulation.”In Proceedings of the International Computer Music Conference. Banff, Canada, 1995, 158–161.

Fitz, Kelly, Lippold Haken, Susanne Lefvert, Corbin Champion, and Mike O’Donnell.“Cell-utes and Flutter-Tongued Cats: Sound Morphing Using Loris and the ReassignedBandwidth-Enhanced Model.” The Computer Music Journal 27, 4: (2003) 44–65.

Fitz, Kelly, and Lippopld Haken. “On the Use of Time-Frequency Reassignment in AdditiveSound Modeling.” Journal of the Audio Engineering Society 50, 11: (2002) 879–892.

Garton, Brad. “Multi-language Max/MSP.”, 2007, Accessed August 27, 2008. http:

//www.cycling74.com/story/2007/3/26/145938/572.

Grey, John M. “Multidimensional perceptual scaling of musical timbres.” Journal of theAcoustical Society of America 61, 5: (1977) 1270–1277.

Grisey, Gérard. “Did You Say Spectral?” Contemporary Music Review 19, 3: (2000) 1–39.

Harris, Fredric J. “On the Use of Windows for Harmonic Analysis with the Discrete FourierTransform.” Proceedings of the IEEE 66, 1: (1978) 51–83.

Harvey, Jonathan. In Quest of Spirit: Thoughts on Music. Berkeley, CA: University ofCalifornia Press, 1999.

Lagrange, Mathieu, Sylvain Marchand, Martin Raspaud, and Jean-Bernard Rault. “En-hanced Partial Tracking Using Linear Prediction.” In Proceedings of the 6th InternationalConference on Digital Audio Effects. London, UK: DAFx-03, 2003, 402–405.

Lagrange, Mathieu, Sylvain Marchand, and Jean-Bernard Rault. “Partial Tracking basedon Future Trajectories Exploration.” In Proceedings of the 116th Audio Engineering SocietyConvention. AES, 2004.

. “Enhancing the Tracking of Partials for the Modeling of Polyphonic Sounds.” IEEETransactions on Audio, Speech, and Signal Processing 15, 5: (2007) 1625–1634.

Laroche, Jean. “Synthesis of Sinusoids via Non-Overlapping Inverse Fourier Transform.”IEEE Transactions on Speech and Audio Processing 8, 4: (2000) 471–477.

Page 113: Klingbeil Dissertation Web

References 103

Laroche, Jean, and Mark Dolson. “Phase-vocoder: About this phasiness business.” InIEEE ASSP Workshop on Applications of Signal Processing to Audio and Acoustics. IEEE, 1997,19–22.

. “New Phase-Vocoder Techniques for Pitch-Shifting, Harmonizing and Other ExoticEffects.” In IEEE ASSP Workshop on Applications of Signal Processing to Audio and Acoustics.IEEE, 1999, 91–94.

Lerdahl, Fred. “Timbral Hierarchies.” Contemporary Music Review 1, 2: (1987) 135–160.

Levine, Scott N. Audio Representations for Data Compression and Compressed Domain Processing.Ph.D. thesis, Stanford University, Stanford, CA, 1998.

Lindemann, Eric. “Music Synthesis with Reconstructive Phrase Modeling.” IEEE SignalProcessing Magazine 24, 2: (2007) 80–91.

Loy, Gareth. “Composing with Computers—a Survery of Some Compositional Formalismsand Music Programming Languages.” In Current Directions in Computer Music Research,edited by Max V. Matthews, and John R. Pierce, Cambridge, MA: The MIT Press, 1989,chapter 21, 291–396.

Lyon, Eric. “Spectral Tuning.” In Proceedings of the International Computer Music Conference.Miami, FL: ICMC, 2004, 375–377.

Maher, Robert C. An Approach for the Separation of Voices in Composite Musical Signals. Ph.D.thesis, University of Illinois, Urbana, IL, 1989.

Matthews, Max V. “The Digital Computer as a Musical Instrument.” Science 142, 3592:(1963) 553–557.

McAulay, Robert J., and Thomas F. Quatieri. “Speech Analysis/Synthesis Based on ASinusoidal Representation.” IEEE Transactions on Acoustics, Speech, and Signal Processing34, 4: (1986) 744–754.

Misra, Ananya, Perry R. Cook, and Ge Wang. “Musical Tapestry: Re-composing NaturalSounds.” In Proceedings of the International Computer Music Conference. ICMC, 2006.

Murail, Tristan. “Spectra and Pixies.” Contemporary Music Review 1, 1: (1984) 157–170.

. Pour adoucir le cours du temps. Editions Henry Lemoine, 2005.

Nuttall, Albert H. “Some Windows with Very Good Sidelobe Behavior.” IEEE Transactionson Acoustics, Speech, and Signal Processing 29, 1: (1981) 84–91.

Page 114: Klingbeil Dissertation Web

References 104

Pampin, Juan. “ATS: A System for Sound Analysis Transformation and Synthesis Basedon a Sinusoidal plus Crtitical-Band Noise Model and Psychoacoustics.” In Proceedings ofthe International Computer Music Conference. Miami, FL: ICMC, 2004, 402–405.

Phillips, Andrew T. “A Containter for a Set of Ranges.” Dr. Dobb’s Journal 17, 6: (1999)75–80. http://www.ddj.com/cpp/184403660?pgno=1.

Plante, F., G. Meyer, and W. A. Ainsworth. “Improvement of Speech Spectrogram Accuracyby the Method of Reassignment.” IEEE Transactions on Speech and Audio Processing 6, 3:(1998) 282–286.

Puckette, Miller. “Phase-locked vocoder.” In IEEE ASSP Workshop on Applications of SignalProcessing to Audio and Acoustics. IEEE, 1995, 222–225.

. Theory and Techniques of Electronic Music. World Scientific Press, 2007. http:

//crca.ucsd.edu/~msp/techniques/v0.11/book.pdf. Online edition draft, December30, 2006.

Roads, Curtis. Microsound. Cambridge, MA: The MIT Press, 2001.

Robel, Axel. “A New Approach to Transient Preservation in the Phase Vocoder.” InProceedings of the 6th International Conference on Digital Audio Effects. London, UK: DAFx-03, 2003.

Robel, Axel, and Xavier Rodet. “Efficient Spectral Envelope Estimation and its Applicationto Pitch Shifting and Envelope Preservation.” In Proceedings of the 8th InternationalConference on Digital Audio Effects. DAFx-05, 2005.

Rodet, Xavier, and Philippe Depalle. “Spectral Envelopes and Inverse FFT Synthesis.” InProceedings of the 93rd Audio Engineering Society Convention. AES, 1992. Preprint no. 3393.

Rodet, Xavier, and Adrien Lefèvre. “The Diphone program: New features, new synthesismethods, and experience of musical use.” In Proceedings of the International ComputerMusic Conference. Thessaloniki, Grece: ICMC, 1997, 418–421.

Sandowsky, George. “My 2nd Computer was a UNIVAC I: Some Reflections on a Careerin Progress.” Connect: Information Technology at NYU http://www.nyu.edu/its/pubs/

connect/archives/index01.html.

Schoenberg, Arnold. Theory of Harmony (Harmonielehre). Translated by Roy E. Carter.Berkeley, CA: University of California Press, 1983. First published 1922.

Schwarz, Diemo. Spectral Envelopes in Sound Analysis and Synthesis. Master’s thesis, Institutfür Informatik, Stuttgart / IRCAM, 1998.

Page 115: Klingbeil Dissertation Web

References 105

. “Current Research in Concatenative Sound Synthesis.” In Proceedings of theInternational Computer Music Conference. Barcelona, Spain: ICMC, 2005.

Serra, Xavier. A System for Sound Analysis/Transformation/Synthesis based on a Deterministicplus Stochastic Decomposition. Ph.D. thesis, Stanford University, Stanford, CA, 1989.

Shneiderman, Ben. “Direct Manipulation: A Step Beyond Programming Languages.” IEEEComputer 16, 8: (1983) 57–69.

Smith, Julius O. “Viewpoints on the History of Digital Synthesis.”, 1992, Accessed July 52007. http://ccrma.stanford.edu/~jos/kna/kna.pdf.

Smith, Julius O. “Virtual Acoustic Musical Instruments: Review and Update.” Journal ofNew Music Research 33, 3: (2004) 283–304.

Smith, Julius O., and Xavier Serra. “PARSHL: An Analysis/Synthesis Program for Non-Harmonic Sounds Based on a Sinusoidal Representation.” In Proceedings of the In-ternational Computer Music Conference. Champaign-Urbana, IL: ICMC, 1987, 290–297.http://ccrma.stanford.edu/~jos/parshl/parshl.pdf.

Stilson, Timothy. Efficiently-Variable Non-Oversampled Algorithms in Virtual-Analog MusicSynthesis. Ph.D. thesis, Stanford University, Stanford, CA, 2006. http://ccrma.stanford.edu/~stilti/papers/TimStilsonPhDThesis2006.pdf.

Taube, Heinrich. “λGTK: A Portable Graphics Layer for Common Music.” In Proceedingsof the International Computer Music Conference. Barcelona, Spain: ICMC, 2005.

Terhardt, Ernst, Gerhard Stoll, and Manfred Seewann. “Pitch of complex signals accordingto virtual-pitch theory: Tests, examples, and predictions.” Journal of the Acoustical Societyof America 71, 3: (1982) 671–678.

Välimäki, Vesa, and Antti Huovilainen. “Oscillator and Filter Algorithms for VirtualAnalog Synthesis.” Computer Music Journal 20, 2: (2006) 19–31.

Varèse, Edgard, and Chou Wen-chung. “The Liberation of Sound.” Perspectives of NewMusic 5, 1: (1966) 11–19.

Wishart, Trevor. “Computer Sound Transformation: A personal perspective from the U.K.”,2000, Accessed August 26, 2008. http://www.composersdesktop.com/trnsform.htm.

Wright, Matthew, Amar Chaudhary, Adrian Freed, Sami Khoury, and David Wessel. “Au-dio Applications of the Sound Description Interchange Format Standard.” In Proceedingsof the 107th Audio Engineering Society Convention. AES, 1999a. Preprint no. 5032.

Page 116: Klingbeil Dissertation Web

References 106

Wright, Matthew, Richard Dudas, Sami Khoury, Raymond Wang, and David Zicarelli.“Supporting the Sound Description Interchange Format in the Max/MSP Environment.”In Proceedings of the International Computer Music Conference. ICMC, 1999b.

Zbyszynski, Michael, Matthew Wright, and Edmund Campion. “Design and Implemen-tation of CNMAT’s Pedagogical Software.” In Proceedings of the International ComputerMusic Conference. ICMC, 2007.

Zeitlin, Vadim. “The wxWindows Cross-Platform Framework.” Dr. Dobb’s Journal 24, 3:(2001) 106–112.

Page 117: Klingbeil Dissertation Web

107

Appendix A. Tear of the Clouds score

Tear of the Clouds is a composition for 13 instruments and electronic sounds. It was first

performed May 6, 2008 by the Manhattan Sinfonietta, Jeffrey Milarsky, conductor and

music director. In performance, the ensemble includes a MIDI keyboard that functions in a

dual role: as a traditional keyboard that controls a variety of sampled and quasi-synthetic

pitch sounds, and as a trigger for pre-synthesized sonic complexes. The tuning of the

keyboard is dynamic and adjusts to the prevailing non-tempered harmony. Pre-synthesized

sonic events were realized using a combination of SPEAR, Common Music, and CLM.

Realtime synthesis and sample playback was done using SuperCollider.

A.1 Program Note

As the highest lake in the Adirondacks, Lake Tear of the Clouds is the Hudson River’s

source. The piece was inspired in large part by my experiences and observations living

along this waterway — the vast wide expanse at Tappan Zee, the churning gray whitecaps

of a stormy day, the vistas from the New Jersey Palisades, and an imaginary slow motion

dive from the top of the cliffs into the murky depths. Some of the sonic materials used in

both the ensemble and electronics are derived from recordings of water sounds made at

various points along the river. Other sound spectra are derived from sources such as bells,

tam-tams, and voices. In many instances these inner structure of these sounds are retuned

via a process of analysis and resynthesis.

Page 118: Klingbeil Dissertation Web

Rev 3/16/08

Tear of the Clouds

Instrumentation

Flute doubling PiccoloOboeBb Clarinet doubling Bb Bass ClarinetBassoon

HornTrumpetTrombone (with F trigger)

Percussion crotales (upper octave, written C5–C6) vibraphone (with motor) marimba (4 1/3 octave, A3–C7) temple blocks (set of 5) sandpaper blocks (one block mounted, the other free and playable with one hand) Chinese cymbal medium suspended cymbal 2 bongos 2 conga drums, high and low (or toms) large tam-tam bass drum

Computer / electronic sounds (see explanation below)Keyboard 88-key MIDI keyboard with sustain pedal, connected to computer

ViolinViolaCelloContrbass

All instruments sound the written pitch with the following standard exceptions:

Piccolo — 1 octave higher Crotales — 2 octaves higher Contrabass — 1 octave lower

B

Bb

µ

Â

Electronics

congas

bass dr. bongos tam-tam

sus. cym.

Chinesecym.

temple blocks

quarter-tone flat

three-quarter-tones flat

quarter-tone sharp

three-quarter-tones sharp

Accidentals Strings

exaggerated bow pressure creating a low, noisy, “crushed” sound

resume normal bow pressure

Note: strings should play with ordinary vibrato unless otherwise indicated

There are two types of electronic sounds indicated in the score, those than are triggered by the MIDI keyboard and those that are triggered by the computer.

The MIDI keyboard functions much like a traditional keyboard playing a variety of sampled and quasi-synthetic pitched sounds. A timbre description is indicated in quotation marks (e.g. “bowed pluck”). The computer operator takes care of switching timbres at the appropriate time. The actual pitches produced, which are often microtonally inflected, are indicated in the staves labeled “Synthesis.”

The computer operator also triggers various sonic events — chords, sustained tones, noise complexes, etc. These are also generally indicated in the “Synthesis” staves.

Whenever possible, woodwinds should execute quarter tones using alternate fingerings. Trumpet and horn should use alternate fingerings based on the 7th and 11th partials.

sandpaperblocks

Duration — 10'54"

A. Tear of the Clouds score 108

Page 119: Klingbeil Dissertation Web

&

&

&

?

&

&

?

&

&

15

&

?

&

&

15

&

?

&

B

B

?

4

3

4

3

4

3

4

3

4

4

4

4

4

4

4

4

4

3

4

3

4

3

4

3

4

4

4

4

4

4

4

4

4

3

4

3

4

3

4

3

4

2

4

2

4

2

4

2

Flute

Oboe

Bb Clarinet

Bassoon

Horn

C Trumpet

Trombone

Computer

Violin

Viola

Cello

Contrabass

Keyboard

Synthesis

Percussion

q = 66

q = 66

˙

F

°

Œ

vibraphonearco

(lock pedal down)

c1

.

.

.

.

.

.

.

J

œœ

œ

œœ

œœµ

B

b

m

B

f

“big pluck”

œ

f

.œmÓ

o

p

œn>

f

.ϵ>

˙

p

Ó Œ ‰

J

œm>

Í

˙µ

>

Í

˙

˙

Ó

c6

œ

œœœ

µ

m

m

p

œ

w

o

F

∑&

w

Ó ˙

>o

Í

(soundingas written)

sul G

&

œ œ œ

F

œ

Œ ‰J

œ>

Í

œ

Ó ⋲

.

J

œ

π

con sord.

Œ

˙

c7

Œ

œ

œœ

œœ

œ

B

Œ

ϵ

œ

o

Œ

ø

œB>

f

5

Œ

J

œÂ>

Í

˙

3

o

F

˙ œ

π

œ ˙

F

˙ œ

F

˙

p

œ

˙ œ

˙

F

œ

œ

o

œ

o

gliss. harm.

œ

o

F

wB

π

wb

π

œ

J

œ

ø

‰ Ó

.˙B>

f

con sord.

J

œ

ø

˙

f

Ó

œ

œœœœ

B

µm

µ

µ

>c1

œœB

p

œœœ

œœ

mb

F

Œ Ó

“bells”

wb

π

wB

p

w

p

r

œ

o

f

‰ . Œ Ó

w

F

w

F

take Bass Clarinet

w

w

w

œ

Ó

ø

œ

Ó

ø

œ

Ó

ø

œ

Ó

ø

œ

Ó

ø

3/16/08

Tear of the CloudsMichael KLINGBEIL

© Copyright 2008 Michael Klingbeil

C SCORE

A. Tear of the Clouds score 109

Page 120: Klingbeil Dissertation Web

&

&

&

?

&

&

?

&

&

15

&

?

&

&

15

&

?

&

&

B

&

4

2

4

2

4

2

4

2

4

4

4

4

4

4

4

4

4

2

4

2

4

2

4

2

4

5

4

5

4

5

4

5

4

3

4

3

4

3

4

3

Fl.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~

8

8

8

8

∑B

8

8

8

8

mediumyarn

÷

8

8

8

q = 84

q = 84

wm

p

w

w

p

w

Œ

.˙B

π

œ˚

π

J

œ

π

˙

Chinese cym. sus.cym.

œœ

œœB

µ

B eeeeeeeeeeeeeeeeeeeeeeeeeœ

œbB

p

p

∑&

wmw

w w

˙

P

œ

R

œ

p

‰ .

w

wÂwµ

wµwÂ

⋲ .

J

œ

ø

œBœ

œBœ

œBœ

œBœ

œBœ

œBœ

F

œB

3 7

Œ œ

ø

œ œ

œmœ

⋲œ

œµœ

œµœ

6

œ œ

F

œ œÂœµ

œ

ϵ

5

œbB

p

J

œn˘

f

œ

P

j

œ

fl

F

low conga

œÂœµ

F

œ

œ

œB œ

ϵ

6

œµœÂ

F

j

œ

pizz.

f

.œ.œB

p

?

p

œ

ƒ

œœµ œ

œB.œµ œ

p

J

œÂ

p

˙

œ

ƒ

œ

œÂœm œB .œ

p

J

œn ˙

6

Ó ‰

œm

p

œ œ

F

œ

3

Ó ‰

œ

p

con sord.

œ œ

F

œ

3

Ó ‰ œb

p

œ œ

F

œ

3

(con sord.)

J

œ

˘

‰ ⋲

.

j

œ

p

j

œ‰ Ó

low bongo

c2

œœ

œ

œœœœ

œ

œ

B

B

µ

µ

mB

∑œ

p

œ

œ

œ œ

œm œ

f

œ œ

Ó .

ϵ

f

ϵ

œµœm œB

œm œ œ

j

œJ

O

œn

p

O

˙

7

œ

œÂ

µ

P

arcoœ

œ

œ

œ

f

œ

œ

‰ .

R

O

œ

BB

p

O

œ

3

œ

œm

P

œ

œ

œ

œ

f

œ

œ ‰

J

O

œ

mm

p

O

œ

3

œm

æ

œ?

2 Tear of the Clouds

3/16/08

A. Tear of the Clouds score 110

Page 121: Klingbeil Dissertation Web

&

&

&

B

&

&

?

÷

&

15

&

?

&

&

15

&

?

&

&

?

?

4

3

4

3

4

3

4

3

4

4

4

4

4

4

4

4

4

2

4

2

4

2

4

2

4

4

4

4

4

4

4

4

4

3

4

3

4

3

4

3

4

4

4

4

4

4

4

4

Fl.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~

13

˙

æ

F

flz.

.

J

œb

æ

13

13

∑?

13˙

J

œ

ø

‰?

13

˙

j

œ

ø

13

˙

J

œ

ø

13

13

13

13

13O

œ

O

œ

J

J

O

œ

f

O

œ

O

œ

J

J

O

œm

f

O

œ

O

œ

J

J

œ

œ

BB

f

œ

f

A

A

œm

ƒ

Œ Ó

jetwhistle

œ œ œµœm œn

œmj

œ

fl

ƒ

‰ Ó

6

Bass Clarinet

œµœm œn

œ œm œb

j

œ

fl

ƒ

‰ Ó

6

˙

>Sf

senza sord.

j

œ

p

‰ Œ

j

œ

fl

f

‰ Œ Ó

bass drum

œœœµ

µB

c3 / c0

œ

œœœ

B

œ

œ

œ

œ

Â

m

m

ƒ

voice

“big pluck”

ƒ

°

Œ

J

O

œ

‰ Œ Ó

J

O

œ

‰ Œ Ó

J

œ

œ ‰ Œ Ó

œ

ƒ

œ

F

œ ˙

q = 60

q = 60

˙

f

Ó Œ ‰

j

ϵ

ø

Œ

.˙m

ø

senza sord.

Ó ⋲

.

j

ϵ

ø

œ

w

˙ œ œ

F

Œ

˙

ø

œ œ

F

œ œ

Œ ˙

ø

senza sord.

π

œ œ œ

F

œ

Ó ‰

j

œÂ

ø

ø

F

œ œ œ

F

œ

Œ

˙m

ø

œ ˙

F

œ

p

˙

3Tear of the Clouds

3/16/08

A. Tear of the Clouds score 111

Page 122: Klingbeil Dissertation Web

&

&

?

?

&

&

?

÷

&

15

&

?

&

&

15

&

?

&

&

?

?

4

4

4

4

4

4

4

4

4

2

4

2

4

2

4

2

4

4

4

4

4

4

4

4

Fl.

Ob.

Bs. Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

19

œ Œ

ø

˙B

ø

19

J

œ

ø

˙µ

ø

œ œ

F

19

œ .˙

F

19

wb

>

π

19

‰ j

œbB

ø

˙ œ

F

œ

19

œ

ø

J

œB

ø

˙

19

wb

>

f

19

œ

f

Œ Ó

tam-tam

19 c1

œb

19

œb

>

f

Œ Ó

19

Ó

˙

˙

b

ø

œ œm

π

œ

œÂ

˙

˙

w

f

w

>

ø

ø

ø

˙

˙

˙

˙

˙

˙

F

˙

˙

˙

F

œ

œ

œ

œ

F

˙

˙

w

F

œ

ø

3

Œ

.˙Â

ø

w

Πj

œ

ø

3

œ

ø

J

œ

ø

˙

r

œ

ø

‰ .

c2

œB

œ

>

Œ Ó

œ

œ

.

.

˙

˙

m

π

œ

œ

œ

œ

œ

œ

µ

π

˙

˙

w

w

Â

f

wB

>

ƒ

œ

J

œb

p

3

œ œ

.œm

F

œ .œ œµ

œ

œ

f

˙ .œ œm

œ œb

ƒ

œ œ

j

œbB

F

œ .œ œB

3

œ œ

F

˙ .œ œ

‰j

ϵ

ƒ

˙ œ

c3 c4

‰j

ϵ

Œ Œ ‰ .r

œ

‰j

œm

>

Œ Œ ‰ .r

œ

œ

œ

œ

P

˙ .œ œm

.

œ

œ

œ

µ

F

˙

˙

.

.

œ

œ

œ

œB

µ

B

‰j

ϵ

ƒ

.˙ .œ œ

>

˙

J

œ œn œ

ϵ>

f

3

˙ œ.œ œ œm

>

f

.˙ œ œ>

.˙ œ œ

p

œ œ

j

œ .œ

j

œ

>

f

3

˙ .œB

J

œ

>

f

˙ ˙m

>

c5 c6

Ó j

œm

Œj

œ

Ó j

œm

Œj

œ

˙ .œ œB œ œ>

f

.

.

œ

œ

J

œ

œB

µ .

.

œ

œ

J

ϵ

>

f

˙

f

œ œ œ

P

˙ .œ

j

œ

>

4 Tear of the Clouds

3/16/08

A. Tear of the Clouds score 112

Page 123: Klingbeil Dissertation Web

&

&

?

?

&

&

?

÷

&

15

&

?

&

&

15

&

?

&

B

?

?

4

3

4

3

4

3

4

3

8

3

8

3

8

3

8

3

4

4

4

4

4

4

4

4

Fl.

Ob.

Bs. Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

~~~~~~~~~

π

24œ œm œ

œ

œB>

˙

24œ

œ œœb

œ

>

˙

24œ œ œ œ œµ

>œ œ

24

œ .œb

œ

>

f

˙

24

œ œbB œ

œ œbB

>

œ œ

p

24

œ œb œ

œ œ

>

˙

24

œ

⋲ .œÂ

f

œ œ

p

‰ .

24

Œ ‰ . x

>

f

xxxx®

‰ Œ

temple blocks

24 c7 c8 c9

Œ ‰ .

r

œ

œœ

œœœœ

Â

B

µ

µ

‰ j

ϵ

⋲j

œÂ

r

œ

24

Œ ‰ .

R

œœ

œ

b œœ

œŒ

‰ j

œm œ œÂœ œ

Œ

24œ

œB œ œ

œ

µ

m

œ

˚

ƒ

˙

œœµ œ œm

œÂ

˚

ƒ

˙

œ

ϵ

f

œ œ œ

ϵ

p

œ œ œ œ

Ï

q = 76

q = 76

*

œ

p

ϵ

ø

œ œ œ

f

‰ .R

ϵ

ø

œ œ

F

J

œ

ø

‰ Œ Œ

œ

˙µ

ø

œ ˙

f

œm

ø

œ œ

F

œ

˙µ

ø

œ

f

c10

œ

œ

µ

Œ

˙

˙

m

ø

π

⋲⋲ .

œ

BB

ø

œ

œ

J

œ

f

œ

œ œ œ

ø

œ

f

.

.

œ

œ

F

.

œ

f

&

j

œ œ

ƒ

˙>

f

J

œ

œbœm

œÂ

œµœ

Ó ⋲

œ

f

œb

œb œÂœb œn œµ

7

f

Œ

.˙b˚

f

Œ

Ó Œ

œÂ˚

f

x>

ƒ

xxxx‰ Œ Œ

x>

f

xxxx⋲ .

7

c11 c12

œœ

œ

œ

œ

b

B

œœ

œB

b œ

œ

µ

œœ

œ

œ

œ

bg g g g g g g g g g g g g g g g g

ƒ

Ó .

œœ

œ

b

°

Œ Œ

œ

œ

m

* °

˙B>

ƒ

œ

œ

œ

œ

5

Ó .

œ

œ

m

m

.˙>

ƒ

Œ

Ó Œœ

osul G

(soundingas written)

&

œ

œÂ j

œ

>

œµ œ

3

œ

Œ

j

œm

>

œµœm

œm œœµ œ

3

4:3

Œ ŒJ

ϵ>

ƒ

œ œ œ œµ œ œb

33

Œ

œm˚

ƒ

Ó

Œ

œ

>

J

œbB˚

œ œ

3

Œ œ

>

J

œ

˚

œ œ

3

œ

Œ

J

œB

˚

œ œ

3

Œ Œx

>

x x x x‰ œ

p

6 3sus. cym.

c13 c14 c15

Œœ

Œ

œm

j

œ

‰ .

R

œÂ

3

Œ

œ

œm

°*

j

œ

‰ .R

œ

* °

œ

3

œ

œ

ϵ

J

œµ .œ

œ

œ

œ

œ

3

œ

œ Œj

œµ œ

œ œmœµ

œmœ œ

3

5:3œ

r

Œ œBj

œ

œ œBœ

œ

33

&

œ

o

Œj

œ œ

Œ

3?

5Tear of the Clouds

3/16/08

A. Tear of the Clouds score 113

Page 124: Klingbeil Dissertation Web

&

&

?

?

&

&

?

÷

&

15

&

?

&

&

15

&

?

&

&

&

?

4

2

4

2

4

2

4

2

4

3

4

3

4

3

4

3

4

4

4

4

4

4

4

4

Fl.

Ob.

Bs. Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

~~~~~~~~~~~~

p

29œ

œÂœ œm

œb œn œm œ˙b

Ï

3

29 œ

.œ œ

œ

œµœ

œ

œ œm˙b

Ï

7

29

œµ œn.œ œ

œ œ œ

œ œ ˙

9

29

29r

œ‰ . ‰

œ

>

˙bB

ƒ

3

29

⋲ .

J

œ œ.œb

œb

˙n

ƒ3

29

Œ ‰

œ>

f

˙

3

29

œ

J

œ

j

œ œ

>

f

œ>

Œ

3 Chinese cym.hi conga

29œœ

œœ

œœ

œ

µ

m

m

µ

Œ

c16 c17

c18

Œ ‰

œ

3

.

j

ϵ

29

Ó

œ

œmœ

œœ

œm œ

Œ

7

.

j

œ

œm œœb

œb œm

5

29

⋲ .

J

œm œ œm œµ œmœ œb

˙µ

æ

ƒ

6

⋲.

J

œb œ ˙µ

æ

Ï

.

j

ϵ

j

œ

œÂ

˙

æ

ƒ

3

*

˙

˙

œ

œB˘

Ï

œm˘

œ

˘

œm˘

œb

˘

5

Œœm

fl

f

œ

fl

œb

fl

œm

flœ

fl

5

˙

˙

t

t

∑t

˙µ

˙µ

˙

q = 69

q = 69

B

B

œ

Œ

œ

Œ

œ

>

œ œ

œmœm

œ œm

œ œb

>

œ œb

3

˙

>

Ï

œ Œ

œ

Œ

œb

>

f

œ

p

Ͼ

S

Ͼ

low conga

Œœb

Œœb

“bowed pluck”

Œ

œb

f

°

.

J

œ

⋲ Œ

.

J

œ

œ

˝

pizz.

ƒ

B

.

J

œ⋲

œm

˝

pizz.

ƒ

?

œ

>

f

œœm>

ϵ>

œm

>

ϵ

>

5

œb

π

œb œn œb œb œn œµ

œ

œb œn œb œb œn œµ

œ œ œb œb œn œµœ

7 7 7

˙

j

œ

ø

œæ Ó

π

c1

‰ .R

œ

p

‰ .r

œB œ œ

‰ .R

œ

p

‰ .r

œ œ œ

œ

>œÂ

>

œ

>

œ ˙

p

3

6 Tear of the Clouds

3/16/08

A. Tear of the Clouds score 114

Page 125: Klingbeil Dissertation Web

&

&

?

?

&

&

?

÷

&

15

t

t

&

&

15

&

t

&

B

?

?

4

4

4

4

4

4

4

4

Fl.

Ob.

Bs. Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~

Ÿ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~

33

33

33

œ œœ

œµ œ

œm

œb

œn

œb œ œbœb œb œb

œb œb

6

7:6

33

Ӝ

π

. .œ

œb

33

33

33

Œ

ø

33

33

œ

œÂœµ

œB œ

ϵ

ϵ3

33

œ

œœm

œ œ

œ

œm3

33

Ó Œ ⋲

.

j

ϵ

ø

arco

‰ j

œb

arco

ø

Œ

ø

œ œµœ œ

œb œb œn œµœ

œb.œb œb œb

œb œn œb œb œn œµ

œ œ œb œb

7 76 5

J

œ œ

œµ œb

3

Œ

ø

tam-tamB.D. beater

œ

.œb

œm . .œ œÂ œœµ

œb œ œµœµ œm

3 37

œ

.œb

œm . .œ œn

* °

œ

œmœb œ œm œ

œm

3 37

œ œ

F

˙

Œ

ø

F

j

œ

ø

J

œm

ø

œ œ œ

F

œ

œ œ

œ

œb

œm

œœb

œbœb

œb œn œb œb œn

3 3

.œm œm œ

œ

œµ œœm œm œ .œb

œ5

5

55

w

p

l.v.

œ

œ œµœb œ œµ

œµœb

œµœµ œm œµ œ

œµœB

œb

j

œ

Œ

6 6

œ

œ œmœb œ œm

œmœb

œm œœm œm œ

œmœ

œb

j

œ

Œ

6 6

Œ ‰œµ( )

j

œ

sul pont.

π

˙

œŒ

ø

Ó

7Tear of the Clouds

3/16/08

A. Tear of the Clouds score 115

Page 126: Klingbeil Dissertation Web

&

&

?

?

&

&

?

÷

&

15

t

t

&

&

15

&

t

&

B

?

?

Fl.

Ob.

Bs. Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~

~~~~~~~~Ÿ~~~~~~~~~~~~~~~~~~~~~~~~~~~~

Ÿ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~

36

36

36

œµœ

œb œœb œb œb œn œµ

œ

œb œn œµœ

œ

œB

p

œ

œ

œb

37

7

36

œ

œ œ œbœm œm œb

œn œmœm

p

œ

œµ œn

œ

œµ œb

5

6:5

36

36

36

36

Ó Œ ‰

J

œ

F

marimbamediummallets

?

36

œB

œ

‰ œµ

œ

œmœb œµ œµ œ œ

.œB

5 3

5

?

36

œ

œ

‰ œm

œ

œmœb œm œm œ œ

5 3

5

?

36

œ

.ϵ

P

œb

œB

ord.

.œŒ

5

Ó ⋲

.

J

œ

P

œ

ϵϵ

3

œ œ œbœb œb œb œn œµ

œ œ

œb

33

5

œm

œ

œ œ œbœm œn œm œb

œmœb œm œ œm

œb œn œ

œ

œn œmœb œn œ

œ

œ

5

6

7

7

œ .˙

p

wb

ø

‰ . .rK

œµ .œ

œ œ

œ œB œ œµ œ œ œµ

5 5

‰ . .rK

œm

* °

œ œ

œ œ œ œm œ œ œm

5 5

œ( )˙µ

p

œœb œ

œb

œ œµ œµ

3

œ œb

œµœµ œ œ œm œm œn

œB œ .œ

5

œb

poco cresc…

œb œb œµœ

5

œ

œ

œb

œœm œ

œ œm œ œb œ

œm

œb œ

œmœµ

5

3

˙ .œ œ˘

F

œ œ œbœb

œ œb

6

w

œµ œbœµ œm

ϵ

œ

œœ œµ œm œn œ œb œµ œµ œ

œb œµ œµœB

73 5

œm œbœm œm

œm

œ

œœ œ œ œn œ œb œm œm œ

œb œ œ

œ

73 5

œœµ

œœm

F

œ( ).˙m

sul pont.

.œB

f

œ .

j

œ œ œ

œm .œœ

P

œb œmœµ5 7

8 Tear of the Clouds

3/16/08

A. Tear of the Clouds score 116

Page 127: Klingbeil Dissertation Web

&

&

?

?

&

&

?

?

&

15

t

?

&

&

15

&

?

&

B

?

?

Fl.

Ob.

Bs. Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

Ÿ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~

~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~

~~~~~~~~~~~~~~~~~~

39

39

39

œ

œ œ œ

œ

œB œ

œ

œb

3

39œ

œ

œnœm

œ

œ

œn œ

cresc…

œ œ œ œbœm œn œ œb œ

3

55

39

39

39

39

œ œ#

œœb œ

œ

œ œb

œœ

p

œ#

5 3

39

w

‰ .R

œ œ œ.œµ .œ

œ

39

‰ .R

œ

* °

œ œ.œm .œ

œ

39

‰ .r

ϵ

Í

œ œœ œ

.œB

5

˙

ord.

Ó

ø

œbœµ

œ

œÂ œ

J

œ

p

‰ Œ

œ œœ

P

œ œ œb œ œb œ œbœb œµ

œ

œ

œ

œµ œn œ

6

œµ œ

œœ œ

œn œ œbœm œn œm œb

œm

f

6

6

‰ J

œ

‰J

œb

‰ .

r

œb

π

Œ

w

œ

œb

ϵϵ

œb

œ

5t

œ

œbœ

œm

œb

œ

5t

œœb

œÂ œ

œBœ

œBœ

œnœµ œµ

œµ œbœ

œBœ

35

3

5

ø

ø

œ

f

œ

œb

p subito

œ

œb

.œb .œ œb

œ

p subito

œ œ œmœ œb œm œm

œn œmœ œb œm œm

œmœb

œœm œm

œmœb œ œÂ

œ6

6

66

œœ

f

.

.˙˙

p

c2 / 12

w

π

p

œm

?

œm

* °

?

œb( )wµ

f

sul pont.

Œ

.˙Â

ø

sul pont.

œ

F

pizz.

Œ ⋲

.

j

œµŒ

9Tear of the Clouds

3/16/08

A. Tear of the Clouds score 117

Page 128: Klingbeil Dissertation Web

&

&

?

?

&

&

?

?

&

15

t

?

&

&

15

&

?

&

B

?

?

4

3

4

3

4

3

4

3

4

2

4

2

4

2

4

2

4

3

4

3

4

3

4

3

Fl.

Ob.

Bs. Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~

ø

42

42

42

œb œbœ

.œ œb œ.œb œ

3

5 5

42 œ

œm

œb

œm

œ

œmœm

œœÂ œb œb

œb œb œnœ œb

5

42

42

42

42

42

w

œB œb œb œBœ

5

42

œ œb œb œœ

5

42

Ó

˙

˙

µ

µ

ø

ord.

œ

F

Œ

J

œ

œb

J

œµŒ

3

‰ J

œÂ

ø

Œ.˙b

ø

œœ

œb

œ

œb

œb œœ œb

œbœb

œB.œ3

œ œœ œ œ

Œ

Ó

˙bB

ø

∑ ÷

w

œb˙

œn œbœm œb œB

3

5

t

œb˙

œn œbœm œb œ

3

5

t

w

wB

ø

˙

˙

œ

œ

p

œ

œ

‰ .R

œb

Œ Œ

r

œ

ϵ

.œm

5

ø

œ œ

p

˙

œ ˙

P

œ

œb

œb

œ œ œmœ

œ œb œb

œ œmœ œb

3

56

œ

P

.œµ œ

œœb

3

œ œ œ

p

œ

.˙Â

ø

con sord.

‰ .r

œ

>

p

˙

π

tam-tam

œµ œœ œbB

œœ

F

˙

6

œm œœ œb

œœ

F

˙

6

œ

œ

p

˙

˙

.

.

˙

˙ &

.

J

œb

f

Ó

*

˙

œ œ

œb œbœb œn

.œb

œbœb

œbœb

37

œbœm

œb

œ

3

˙

˙

p

˙

P

˙

Ó

œ

œ

ø

œ

π

˙

π

10 Tear of the Clouds

3/16/08

A. Tear of the Clouds score 118

Page 129: Klingbeil Dissertation Web

&

&

?

?

&

&

?

÷

&

15

t

t

&

&

15

&

t

&

&

?

?

4

3

4

3

4

3

4

3

4

4

4

4

4

4

4

4

Fl.

Ob.

Bs. Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~

ø

46œ

Ó

ø

46˙

Œ

ø

46œ

œbœ œ œ

ƒ

œ œ

F

œbœ œ

œœ

6

46

Œ

.œm

p

œœm

œœÂ

5

3

46

˙

Œ

ø46

˙m

ø

œ œ

p

46

˙

Œ

ø

46

46 c3 / 20

.

j

ϵ

p

46

“organ pluck”

.

j

œm

p

˙

46œ

˙µ

p

œ O

˙

mm

p

B

Œ

˙B

p

ord.

.

J

ϵ

F

Ó

.˙m

>

Í

˙

Œ

ø

take Bb Clarinet

.œm œ œ

œmœm œ

œœb

˙

j

œ

ø

Ó ‰ j

œb

ø

O

˙

O

œ

∑&

F

.œœ œ œ œm

œ

œm œÂ

œm œ œ œ5

6

B

∑?

˙ œ

p

.

.O

˙

µµ

.

.O

˙

µµ

.

.

O

˙

ÂÂ

o(soundingas written)

sul D

p

Œ ‰

J

œ

p

œ .œ

ϵ

f

Œ

œm-

P

œÂ ˙

œ

π

.

J

œ

⋲ Ó?

Ó Œ ‰ .

r

œ

Í

˙

Œ

ø

‰ .

R

ϵ

Í

Ó . ‰ .r

œb

.˙ .œ

œb

O

˙O

˙

O

˙ O

˙

bb

O

˙

O

˙

˙

o

Œ ‰ .r

œ

osul A

11Tear of the Clouds

3/16/08

A. Tear of the Clouds score 119

Page 130: Klingbeil Dissertation Web

&

&

?

?

?

&

?

÷

&

15

t

t

&

&

15

&

t

&

B

?

&

4

3

4

3

4

3

4

3

4

4

4

4

4

4

4

4

4

3

4

3

4

3

4

3

4

4

4

4

4

4

4

4

Fl.

Ob.

Bs. Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

50

w

æ

p

50

w

f

50

50

œ

p

œœm

œÂœµ

œœm

œÂœµ

œ œ

œÂœµ

œ œ œœ œ œ œµ

6

65

50

œ

50

50.˙

œb

50

50 c4 / 33

50

w

50

O

œ

O

˙

ÂÂ

F

O

œ

p

O

œ

O

˙

F

O

œ

p

J

O

œ ‰ O

˙

bb

F

O

œ

p

w

o

.˙æ

œ ˙µ

F

∑&

œœ œµ œm œ

œm

œm

6

˙

Œ

O

˙

O

œ

µµ

O

˙ O

œ

.

.O

˙

˙æ

Ó

˙ œ œ

f

Ó

œ

p

œµ œµ œœ

œb

œµ œœ

œb œÂ

6

5

Bb Clarinet

œ

p

œµœ

œ

œµœ

œ œÂœ

œ

R

œ

‰ . Œ

6

6

wm

Í

c5 / 40

ϵ

wm

.

.O

˙

F

j

O

œ

π

.

.O

˙

F

j

O

œ

π

Í

∑?

ø

⋲r

œ

O

>

ƒ

‰ Ó

tongue ram

˙œB

œ

‰ .

ϵ

p

œm œ œ

œ œ

6:5

œm

p

œœ

œµœm

R

œ

‰ .

3

3

⋲ .

J

œ

Í

˙

.

j

œ

Í

˙

⋲.

j

œÂ

Í

˙

c6 / 44

.

J

œ

œ.œ ˙

⋲..

J

O

œ

BB

>

f

O

œ

O

œ

.

.

J

O

œm

>

f

O

œ

O

œ

.

.

J

O

œ

bb

>

f

O

˙

p

.

J

œm

pizz.

ƒ

‰ .

R

œ

F

Œ

j

œ

3

12 Tear of the Clouds

3/16/08

A. Tear of the Clouds score 120

Page 131: Klingbeil Dissertation Web

&

&

&

?

?

&

?

÷

&

15

t

t

&

&

15

&

t

&

B

?

?

4

4

4

4

4

4

4

4

4

2

4

2

4

2

4

2

8

5

8

5

8

5

8

5

4

4

4

4

4

4

4

4

4

2

4

2

4

2

4

2

Fl.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

Ÿ~~~~~~~~~~~~~~

54

Œ

.˙m

p

54

œ œm œœÂ

œµ .œœ

ƒ

œ

5

54œb

œ

œb œÂœb .œ

œn

œb œœbB

œbœµ

œ

f

r

œ

‰ .

5

7

54

Ó ‰ ‰ œ

F

œœ

f

œBœµ œµ

œ

36

54

œ Œ

ø

Ó

54

54

˙ Ó

ø

54

54 c7 / 48

Œ œ

?

54

œ

?

54O

œ

P

.

.

O

˙

ÂÂ

ø

O

œ

P

.

.O

˙

BB

ø

O

œ.O

sul D

ø

.

j

œ

p

.O

arco

p

sul D

J

ϵ

F

œn œ

3

˙

œ œµœ œ œ œµ œ

3

˙

O

œ

f

O

œ O

œ

mm

O

œ

f

O

œ

O

œ

b

O

F

OO

œ

µÂ

O

J

O

J

œm

œ œ

J

œ

f

3

J

œ œB.œ

œ

5

J

œm

p

œ

J

œ

J

œ

œ

33

&

œ .œ

.

.O

œ

O

œ

O

œn

>

.

.

O

œ

O

œ

O

œ

µµ

>

.

.O

œ

J

O

œ‰

˙

œÂ.œ œ

œµ œ œBœm œn

œ .˙m

7

Œ ‰

œn( )

j

œb

p

œ œ

œ

f

œb œÂ

3

œ

˙ œ

F

Œ

œ

>

f

.œb

>

˙

c8 / 52

œœ

œœœœœb

B

µ

µ

bŒ Ó

œm

wm

O

wm

ƒ

Œ Œ ‰ . r

œm

sul pont.

Í

œ

Œ

.˙µ

F

œ

pizz.

f

.˙Ko

sul G (soundingas written)

F

arco

&

13Tear of the Clouds

3/16/08

A. Tear of the Clouds score 121

Page 132: Klingbeil Dissertation Web

&

&

&

?

?

&

?

÷

&

15

&

?

&

&

15

&

?

&

B

?

&

4

2

4

2

4

2

4

2

8

5

8

5

8

5

8

5

8

3

8

3

8

3

8

3

4

3

4

3

4

3

4

3

4

2

4

2

4

2

4

2

8

3

8

3

8

3

8

3

Fl.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

p

58˙

58

œ

œmœ œ

œ œm

f

3

58

œœb œb œm œµ œµ œm œµ

œ œœb œ

p

6

6

58

˙

58

58

˙

58

58

58

58

˙

58O

˙

.œ œ

cresc…

œ

œ œb

3

˙&

˙

o

œ œµ œ

œbœ

j

ϵ

5

œ

Π.

œm

F

œœœ

œ

œ

œ

µ

Â

µm

Π.

œ .œ

J

œ

f

œ œµœµ œm œµ

œ œmœÂ

53

œœµ œÂ

œ

f

Π.

&

Œ

.O

sul Dgliss. harm.

f

?

Œ

osul G

gliss. harm.

f

p

œ

>

P

œ

>

œ

j

œ .œ

œµ œm œm

7

j

œ

œ

P

œœœœœœœ

œm

µ

mB

m‰

c9 / 60

.œB

.

.

œ

œ

B

ord.

.Om

f

∑?

œµ œ œœ

œm

ƒ

œm œ

J

œ ˙

œ œ œ œm

Ó ‰

j

œœœ

œœœ

œ

œ

m

b

B

B

µµ

j

œb

˙ œœb

Ó ‰

j

œb

pizz.

f

q = 80

q = 80

œ œ

œb>

ϵ>

œ œ œµ

œœ

œB

f

3

c10 / 68

˙

14 Tear of the Clouds

3/16/08

A. Tear of the Clouds score 122

Page 133: Klingbeil Dissertation Web

&

&

&

?

?

&

?

÷

&

15

&

?

&

&

15

&

?

&

&

?

?

8

3

8

3

8

3

8

3

4

2

4

2

4

2

4

2

4

4

4

4

4

4

4

4

8

7

8

7

8

7

8

7

4

3

4

3

4

3

4

3

4

2

4

2

4

2

4

2

Fl.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

Ÿ~~~~~~~~~~~~~~

Ÿ~~~~~~~~~~~~~

Ÿ~~~~~~~~~~~~~~

~~~~~~~~~~~~~~~

63

f

63 .œ

63

63

63

63.œ

63

.œm

f

63

mediumyarn

63......

œœœœ

œœµ

b

B

c11 / 72 c13 / 80 c14 / 84

.

.œœ

µB

63

t

63

f

.œB

f

.

.

O

œ

µµ

f

.ϵ

arco

p

œ

f

œb

œµœm

3

œœm œm

œ œµ œ

3

Œ ‰

œ

P

œ

œ

3

∑&

œ

>

œ

⋲ .

J

ϵ

P

œ

j

œm

3

œ

æ

>

Í

œ

æ J

œ

>

3

bongos congas

œœ

œ

Bm

c12 / 76

œ

œ

œ

œ

œm

m

b

Â

œœœœœœœ

œ

b

µ

b

µ

n

m

j

œ

œœ

œœœ

µ

m

µ

nn

3

œB

œ

œ

œ

j

œ

œ

3

t

œ

œ

j

œ

3

œ œbœm

œn œµ œ

œB

ƒ

3

œmœb œ œn œ

œ

ƒ

3

œm

f

œnœ

œ œœµ

3

œbœ

j

œb

F

3

p

p

p

p

œ œ

œ˘

œB œœµ>

œ>

œnœb

ϵ

33

˙

Ó

œµœÂ

ƒ

œ

œ œbœ œ œb

œ œmœ

œB ˙7

˙

f

˙

˙

f

˙

œ

j

œœ

F

œ œb œ œ

œm œ œ5

3

˙µ

f

˙

Ͼ

f

j

Ͼ

p

‰ Ó

c15 / 88

œœ

œ

œ

œm

µ

B

œm

wm

f

˙

˙

Ó

˙Ó

˙

Ó

˙

f

˙

.œ œ

Œ

.œbB

ƒ

œ œn˘

.œm

˚

.œ ˙

œ>

F

œ>œ>

‰ ‰ Ó

3

Chinese cym.stick end of mallet

œ

œœ

œ

B

µ

µ

œœ

œm

.œ ˙

.ϵ

œ

Œ

Œœ( )˙

p

Œœb( )˙b

f

Œ ˙bB

f

œ˘

˙b

5

œ

Ó

Œ

˙

p

Chinese cym.with yarn

Œ

œœœ

œœœ

œ

µ

b

m

bb

Œ

œB( )˙m

Í

Œ ˙

˙

æ

Í

Œ

˙µ

Í

15Tear of the Clouds

3/16/08

A. Tear of the Clouds score 123

Page 134: Klingbeil Dissertation Web

&

&

&

?

&

&

?

÷

&

15

&

t

&

&

15

&

t

&

&

?

?

4

2

4

2

4

2

4

2

4

4

4

4

4

4

4

4

4

3

4

3

4

3

4

3

4

4

4

4

4

4

4

4

8

9

8

9

8

9

8

9

Fl.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

Ÿ~~~~~~~~~~~~~~~~~~~~~~~~~~~~

Ÿ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~

Ÿ~~~~~~~~~

~~~~~~~~~~~~~~~~~~~~~~~

~~~~~~~~~~~~~~~~~~~~~~~~

~~~~~~~~~~~~~~~~~~~~~~~

ƒ

ƒ

ƒ

f

68

⋲ .

rK

ϵ

p

œœm

œœm

œ( )œm

68

˙

68

⋲ œœb

œb œB œB

œ

ϵ

œœ

ϵ

66

68

68

œ

j

œ

œ

3

68

F

œ>

ƒ

œ

68

68˙

68

?

68

œ

Œ?

68˙

(with tpt.)

œB

æ

œ

æ

œm

æ

œ>

ƒ

œ

˙µæ

ƒ

ƒ

R

œm˘

ƒ

‰ . Œ Œ ⋲œm

ƒ

œ œÂœµ

œ œµ

7

œÂœ

œ

>

œ œ .œ œ œm

œm

œ

œB œm

f

3

6:5

œ

f

.œ œ

œmœ

œ œ œm

œ

œb œnœm

œm

œmœ

œ

6 5

6

œ

j

œ

œ

j

œ œœ

œB>

ƒ

3 3

p

œ œ

Ï

œ

P

J

œ

œB

J

œ œ

œBœ

f

3 3

P

œœ

œœœ

œ

œ

œœœ

µ

b

m

m

Â

µœœ

œ

œœœn

µ

Bb

c16 / 92 c17 / 96 c18 / 100c19 / 104 c20 / 108

œœ

œ

œœœœ

œ

µ

m

B

b

B‰

œ

œœœœœœ

œ

B

µ

nbBB

œœ

œœœœœ

n

n

Âb

Œ

3 3

œœ

œ

µ

œ

œ

œ

µ

œœ

œ

n

Bœm

œ

33

“big pluck”

œ

j

œ

œ

j

œ

œ

œmœ3

3

ϵ( )wb

Í

w≤

Í

œ

pizz.

‰ œ ‰ œBœm

œ˝

Ï3

3

B

œ

pizz.

f

‰œ

‰œb

‰j

œ

˝

Ï

3

3

Ï

œB>

ƒ

.œ>

˙

˙

Œ

p

œm>

.œ>

˙

˙Œ

ø

Ó œ

F

œbœ

Ó œ

F

q = 76

q = 76

π

π

⋲œm

>

f

œ

>

ϵ>

œ>

.ϵ

œ>

.ϵ

œ>

œm œ

55

œŒ Ó

Œ˙bB

p

œ œB

ϵ

f

.œµ ˙ œ

œb

p

œ

F

Œ Ó

tam-tam

c22 / 116c21 / 112

œ

œœœœµ

b

m

Œ Œ ‰

j

œ

œ

œ

œÂ

B

B

m

.˙ œœb

œ

œmŒ Œ ‰

J

œ

œ

m

F

°

œœb

* °

œ

œ>

f

.œb

œb>

.œ œ

5

B b C

w

>

f

œ( )wµ>

f

16 Tear of the Clouds

3/16/08

A. Tear of the Clouds score 124

Page 135: Klingbeil Dissertation Web

&

&

&

?

&

&

?

÷

&

15

&

?

&

&

15

&

?

&

&

B

?

8

9

8

9

8

9

8

9

4

4

4

4

4

4

4

4

Fl.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

72œb

r

œ œ œµ

œ>

œb

œµ .œ

œm>

œn

3

5

72

72

72

72

˙ œ œb .œ

72

72

˙ œ

œ .œ

72

∑&

72

˙ œ

.œ .œ

72

˙ œ

.œ .œ

t

72

œ

R

œ œ

.œµ œœ

œ

œœ

.œb

p

⋲ œ>

f

5 7

Œ ‰ J

œ

f

œ œ

J

ϵ

Œ ‰ J

œB

f

œ .œ

œ

.œ>

œ

œb

P

œ œœµ œ

5 7

œ Œ

π

˙bB

p

Ó Œ œm

f

œ

Œ

π

Œ

ϵ

f

Œœ

f

°

œbœb

œœ# œ

Ó

7

vibraphone(motor off)

mediummallets

/ 120

Œ

œœ

œœ

œœ

µ

µ

µ

mb Ó

œm

œ

œ

œ

œB

µ

µB

œm

.

.

.

˙

˙

˙n

b

* °

œ œµ œµ œµ œBœm œb

œBœµ

œB ˙5

6

œBœm œb

œBœµ

œB œ

O

œ

>O

œ

3

B

œ

ϵϵ

œBœ

œµ œ œBœ

œœÂ œ

j

œ5

5

3?

Ó ‰

j

œÂ

f

œ

Œ

p

.˙ .

j

œ

ƒ

˙ ‰

œbB>

œbBœb

œm œbœm

œ

5

˙

œB

P

œ

∑?

Œ

p

O

˙Œ

p

œb

f

œ

œµœb

œ

6

œ

˙µ .

j

œ

ƒ

C

C

‰ œµ

>

œ

j

œ

ƒ

œ

œ

F

.œœ

œÂœ œµ .œ œm7

35

.˙m

Œ

f

Œ

π

œ#

F

°*

œ

œbœb œn

œ

Œ Œ ‰ J

œœ

°*

7

/ 124 / 128

œœœ

œœ

b

µÂ

µ

Œ Œ ‰

j

œœ

œœœ

œ

m

B

m

œœ

œB

m‰

j

œ

œ

µ

°*

œ

°*

.˙m

f

œ

œœ

œ

œ

5

.˙ œ

œ

œB

ƒ

œm œµœÂ

œ

œµœµ œm œB œ

œm

œn

7:5

œB

˝

pizz.

Œ Ó

17Tear of the Clouds

3/16/08

A. Tear of the Clouds score 125

Page 136: Klingbeil Dissertation Web

&

&

&

?

&

&

?

&

&

15

&

?

&

&

15

&

?

&

B

?

?

4

3

4

3

4

3

4

3

4

5

4

5

4

5

4

5

4

4

4

4

4

4

4

4

Fl.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

76 œ

œBœ

œb

œ

œn œ œ

œµ œn œœn œ

œÂ

œœm

f

œ

5 66

76

Ó Œ

œ

p

76

œm

>

f

.œ ˙œ

76

76

76

76

76

Ó

œ

œ

#

œ

œ#

j

œ

œ

b

3

76/ 132 / 136 / 140

Ó

œ

œ

œœœœœ

m

µm

œ

œœœœœ

bB

µ

BB

j

œ

œœœœ

µ

nbb

3

Ó

œ

œ

B œ

œ

nm

J

œÂ

3

76

˙œ

*

œmJ

œ

3

76 .˙µ

⋲œm œÂ œ œ

œ œ

7

œ ˙m œm

˙Œ ‰ .

r

œ

f

.œœ œ

œbœb

œ

ϵ

ϵ

œ ˙

œmœ œÂ

œ œ

˙µ

5

œ#

f

°*

.

.

œ

œ

b

j

œ

p

œ.œ œb

5

pedal phrases…

/ 144œœ

œ

œ

œœ

œœ

B

µ

m

B

µ

Ó

œ>

œ

f

°

œ

œ

œbœb

œœm

Ó

7

j

œ œµ œ

j

œ œœ

œm3 3

3

œ

œ œ

.œÂ

œB

˙b œ

œb

F

J

œ

p

Œ ‰

œmœb

œœ

œm œ

>

f

.œB œB œn œ œœb œ

.œm œ

3

œ

f

.œÂ œ.œÂ œ œ

œœb œµ œm

.ϵ

5

5

œ œbœ

œœm

œ .œ

œœm

œœb

J

œ œ œb œBœB

œmœ œb

5

7

œ œ#œ œ

J

œ

f

‰œœ œb

œb.œ

œœn œn

œ

œ œ# œ œn

œ

5

œ œmœb œ œ

œµ œnœµ œµ

œœ œ3 5:3

5:3

œ

œb œ œœµ

œµ œµ œ

5:6

&

œ œ

R

œ .œ œµœ œµ

œµ œ5

53

18 Tear of the Clouds

3/16/08

A. Tear of the Clouds score 126

Page 137: Klingbeil Dissertation Web

&

&

&

?

&

&

?

&

&

15

&

?

&

&

15

&

?

&

&

?

?

4

4

4

4

4

4

4

4

Fl.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

*

79œµ œm œµ œ œm

œ œmœ

œµ œmœ

.œµ œ

‰ .

œœm

œ

œb œ

7:6

5

5

79œ

R

œ œµ œm œµ

œbœ

œ .

j

œB œœ œm

œmœn

J

œm

5

3

79

œœ

œmœ œ

œbœm

œœn

œœm

œn

7 7

79

∑B

79

79

79

79

œb

œœb œ œn œ œ œ# œ

œ

œnœ#

œb

œb

5

79

79

79

œ œm œ œ

œ

œ œ œmœ œ œ

œœn œµ

œ

œb œbœ

œœµ

œ

J

œµ œ œ

R

œ

3

j

:5œ

r

j

:5œ

r

œ œn œ

œ

œœµ œµ

œ

œ

œ

b œnœµ œ

3

5:35:3

B

.œb œµ œ œbœµ

œœµ

œµœb

œ

œbœµ

œbœb

œb

57:5

7:5 3

œ œmœ œb

œ

œÂ œm œm œ œ œmœb œ

œm

œœb

.œ œb œBœB

œmœ œb

œœ

œmœ œ

œb

6:5

Œ

œmœ

œbœ

.œm

f

œn œ

œnœ œ .œ

œ

œ

œ œb œ

œœb œ œn œ œ œ# œ

œ

œn

5

œ

J

œb œ.œ œ

œmœ

œ

œ

œ

r

œ .œ œ œm6:5

5

5

œ œµ .

j

œ

j

œœµ

œœÂ

œœ œb

5:3

5:3

œœm œn

œ

œ œm œ œµ œ

œ

œµœm

9:8

œœ

œm

.œ œœb

œm œmœ œm

œnœ œn

œb œ œ œ

6 6 6

œmœ œ

œmœm

œœn œm

œ.

j

œmœm

œnœ œm

œ.œm

œœ

œœm

œn⋲ œ .œb

.œm .œn

.œ.œm .œ

.œ.œ

6:5 6:5

r

œ œbœ

œm

.œb œ œ

œbœb œ œm

œ

.

j

œ

7:67:6

œ#

œb

œb

œ œœb œn œn œ œ#

œ œ

œœ#

œb

5

œ œ

œ

œ œ œmœ œ œ

œ

œœn œµ

œ œb .œœ

55

5 5

R

œ œm

J

œ œn œnœ

œµœ œ .œ

œ œn

5:3 5:3 5

œbœ

œb œn œœ

œm œnœ

œ.œÂ œµ œ

œ

ϵ

9:8

9:8

19Tear of the Clouds

3/16/08

A. Tear of the Clouds score 127

Page 138: Klingbeil Dissertation Web

&

&

&

B

&

&

?

&

&

15

&

?

&

&

15

&

?

&

&

B

?

4

3

4

3

4

3

4

3

4

4

4

4

4

4

4

4

Fl.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

82 œœ

œmœ

œm⋲

œm

œ œœb

œn œmœ œm

œnœ œn

œb œn.œb

6

6 6

6

82 œ œ œ

œmœ œ

œmœm

œmœm

œnœ œm

œ

6:5 6:5

82œ

œbœ

.œ œ œ

œnœ

œ

œ

œb

œbœm œn

œœm .œ

œ

7:57:5 3

6

82

œ œ œbœ

œmJ

œb œ.œ

œbœb œ œm

œ

œn

7:6 6:5

6:5

82

82

82

82

R

œb

œn œ œœb œ œ

œ#

œb

œ

œ

œbœn œ

œb œn

6

6:5 6:5

82 / 148

œB

82

œB

Œ Ó

82œb

œ

œ

ϵ

œµ .œÂ œn œ

œbœ œ

œ

œnœm

œœ

.œm

55 5

5

œ œµœ

œ œbœ œm œ œm

œnœ

œµœ

œm.œ œb

œœb œn œ

œn

œÂ œ œµ œ

œ

œµœm

R

œn

6:5

6:5 6:5

œœ

œmœ

œmœ œ

œbœn œm

œ œmœn

œ œn

6 35:3 5:3

œ.œ

œmœ œ

œmœm

œmœn

œ œmœ

55

5

œ

œb œbœ

œ œ .œ

œnœ

œ

œ

œb

œœm œ6 6

6

œm œœ

œb œn

œ

J

œ œb

œb

œ œbœb

6:5 6:5

.œ.œ#

œb

œ

œ

œb œœb œn œ

œ#

6:5

5

œn

œb

œ

œ

ϵ

ϵ

œ

œ

œœm œµ œn

œ

J

œm

7:57:5

ϵ

œµœ

œ œbœ œm œ œm

œn œ

ϵ

57:5

J

œ œbœ

œb œn œœ

œ

œ

œµœm

.œn

5

5

5

œb œ.œb œ œ

œœ

œmœ

œm

œ.œm

œœ œ

œb œn.œb .œ

77 7

7

œ.œ

œmœ œ

œmœm

œnœ œm

œ œ œ

œmœ œ

55

55

œ

œœ

œb œbœ

.œ œ

œnœ

œ

œ

œb

œ

œœ

œb œb

6

6

66

R

œ

œm œ œœ

œb œnJ

œ œb

œb

œ œbœb

œ

œœn

œb œn

6:5 55

œb

œ

œ

œb œb œn œ

œ#

œb

œ

œ

œb œ œ

œ#

œb

5

5

55

œn

œb

œ

œ

ϵ

œµœm œµ

œœm œ œn

œb

œ

œ

ϵ

ϵϵ

œœm

6

6 66

œœµ

œœ œb

œ.œm œm

œnœ

œµœ

œµ œbœn œm

R

œ

7:57:5 7:5

.œ œbœ

œb œn œœ

œµœm

.œn œbœ

œb œn œœn œm

œn

55 5 5 5

20 Tear of the Clouds

3/16/08

A. Tear of the Clouds score 128

Page 139: Klingbeil Dissertation Web

&

&

&

B

&

&

?

&

&

15

&

?

&

&

15

&

?

&

&

B

?

4

3

4

3

4

3

4

3

Fl.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

85 œœ

œmœ

œm

œ œœ œ

œb œnœb œ .œ

œœ

œnœ

œm

œ

œb œn.œb

7 7

85œm

œm œ œmœ œ

œmœ œ

œmœm

œmœ

œmœ œ

œm

55 6 7

85 œ.œ œ

œ

œœ

œ

œb

œ

œb œbœ

.œ œnœn

œœ

œ

œb

6

6 3 5:3 5:3

85 œ œb

œb

œ œbœb

œ œ

œn œb

œb

œn œbœb

œ

œn œb

œb

œn œbœb

5 7:5 7:5

85

˙

-

F

œ

-

œ

-

˙

-

3 3

85

85

85

œ

œ

œbœ œ

œ#

œb

œ

œ

œbœ

œ# œ

œb

œ

œ

œb

5 5

5

5

85

85

85 .œ œn

œb

œ

œ

ϵ

œµ œnœ

œm œn

œb

œ

œ

ϵ

œµ œœm œn

œb

6 6 77

œmœ

œœµ

œœµ

œm œ œmœ

œœµ

œœµ œ

œœ

œµœ

œµ œmœ

œ

7:5 7:5 63

œ œbœ

œb œn œœ

œn œbœ

œb œn œœ

œn œbœ

œb œn œbœ

3 6 6 6

œœ

œmœ

œm œb œnœb .œ

œœ

œmœ

œm œnœb œ

œm

œ.œ

œmœ œ

œmœm

œmœ œ

œmœm

œ

œm

7 7 7

œbœ

œ œ œ

œnœ

œ

œ

œb œœ œn

œ

œœ

œ

œb

5:3

7 7

œ

œb

œb

œ œbœb

œ

œb

œb

œn œbœb

œœb

œn œbœb

7:5 7:5

3

R

œ

œ#

œb

œ

œ

œbœ

œ#

œb

œ

œ

œbœ#>

œb

7:5 63

œ

œ

ϵ

œµœm œn

œb

œ

œ

ϵ

œµœ

œb

œ

œ

ϵ

œµ œb

œ

œ

ϵ

ϵ

7 7

œµœ

œµ œœ

œµœ

œµ œœ

œµœ

œµœ

œµœ

œµœ

6 6 6

œbœ

œ œb œbœ

œbœ

œn œb œbœ œ

œn œb œbœ œ

œn

6 3 5:3 5:3

21Tear of the Clouds

3/16/08

A. Tear of the Clouds score 129

Page 140: Klingbeil Dissertation Web

&

&

&

B

&

&

?

&

&

15

&

?

&

&

15

&

?

&

&

B

?

4

4

4

4

4

4

4

4

Fl.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

87œ

œœ

œmœ

œmœb

cresc…

œ œ

œœ

œmœ

œm œ

œœ

œmœ

œmœ

œœm

œ

87

œ œ

œmœm

œ

cresc…

œmœ œ

œmœm œm

œ œ

œmœm œm

œ œ

œmœm

œ

7 7 7

87œ

cresc…

œ

œ

œœ

œ

œb œnœ

œœ

œ

œbœ

œœ

œ

œbœ

œœ

œ

œb

87œ

œb

cresc…

œ œbœb

œœb

œn œbœb

œœb

œn œbœb

œœb

œn œb

6 6

7

87

œ

-

F

œ

-

cresc…

œ

-

87

87

87

œ

œ

œbœ#>

cresc…

œb

œ

œ

œbœ#>

œb

œ

œ

œbœ#>

œb

œ

3 8:5 8:5

87

87

87œb

cresc…

œ

œ

ϵ

ϵ

œ

œ

ϵ

ϵ

œ œµ

ϵ

œm

œ œµ

ϵ

œm

œ œµ

ϵ

œm

œ œµ

ϵ

œµœ

œµœ

cresc…

œµœ

ϵ

œµœ

œµ⋲

ϵ

ϵ

œ

ϵ

ϵ

œ

œ

ϵ

ϵ

œ

6 7

œb œbœ

œn

cresc…

œb œbœ

œn œb œbœ

œb œbœ

œb œbœ

œbœ

œ

œbœ

œ

7

œmœ

œœm

œ

œmœ

œœm

œ

œm

œœm

œ

œm

œœm

œ

œm

œmœ

œm

œmœ

œm

œmœ

œm

œmœ

œm®

œ

œmœm

œ œ

œmœm

œ œ

œmœm

œ

œmœm

œ

œmœm

œ

œmœm

œ

œmœm

œmœm

œmœm

œm

7 7 77

œœ

œ

œb

œœ

œ

œb

œœ

œ

œb

œ

œ

œb

œ

œ

œb

œ

œ

œb

œ

œ

œb

œ

œ

œb

œ

œ

œb

œbœ

œ œbœb

œ

œn œbœb

œ

œn œbœb

œ

œbœb

œ

œbœb

œ

œbœb

œ

œbœb

œ

œbœb

7 7 7

7

œ

-

œ

-

˙

-

œ

-

f5

œ

œb

œb

>

œ

œ

œb

œb

>

œ

œ

œb

œb

>

œ

œ

œb

œ

œb

œ

œb

œ>

œb

œ

œb

œ

œb

8:5 8:5 8:5

œm

œ œµ

ϵ

œm

ϵ

ϵ

œm

ϵ

ϵ

œm

ϵ

ϵ

œm

ϵ

ϵ

œm

ϵ

ϵ

œm

ϵ

ϵ

œm

ϵ

ϵ

œm

ϵ

ϵ

œm

ϵ

ϵ

œm

œ

ϵ

ϵ

œ

œ

ϵ

ϵ

œ

œ

ϵ

ϵ

œ

ϵ

ϵ

œ

ϵ

ϵ

œ

ϵ

ϵ

œ

ϵ

ϵ

œ

ϵ

ϵ

œ

ϵ

ϵ

œ

œµ®

œb

œ

œ

œb

œ

œ

œb

œ

œ

œb

œ

œ

œb

œ

œ

œb

œ

œ

œb

œ

œ

œb

œ

œ œ

œ

œb

æ

?

22 Tear of the Clouds

3/16/08

A. Tear of the Clouds score 130

Page 141: Klingbeil Dissertation Web

&

&

&

B

&

&

?

&

&

15

&

?

&

&

15

&

?

&

&

?

?

4

3

4

3

4

3

4

3

4

5

4

5

4

5

4

5

4

3

4

3

4

3

4

3

4

5

4

5

4

5

4

5

Fl.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

~~~~~~~~~~~~~~~~~~~~~~~

q = 96

q = 96

89 w

ƒ

wm

89wm

ƒ

wm

89 wb

ƒ

wb

89w

ƒ

89

J

œbB

89

w

89 w

89w

ww

w

b

b

æ

æ

ƒ

89

89

89w

w

m

µ

æ

ƒ

w

æ

ƒ

wB

æ

w

f ƒ

.˙.˙m

.˙m .˙m

.˙b.˙b

œ

Ó

ƒ

ƒ

ƒ

.

.

˙

˙

m

µ

æ

.

.

˙

˙µ

æ

B

.˙Bæ

&

.˙æ

œ˘

f

œ˘

œÂ˘

œ˘

œm˘

œ˘

œ˘

œÂ˘

œ˘

œm˘

œ˘

œ˘

œÂ˘

œ˘

œm˘

œ˘

œ˘

œÂ˘

œ˘

œm˘

œ˘

œ˘

œÂ˘

œ˘

œm˘

5 5 5 5 5

œm˚

f

œn˚

ϵ

˚ œm˚ œ

˚

œB˘

f

œ˘

œ

˘ œ˘ œm

˘

œB˘

œ˘

œ

˘ œ˘ œm

˘

œB˘

œ˘

œ

˘ œ˘ œm

˘

œB˘

œ˘

œ

˘ œ˘ œm

˘

œB˘

œ˘

œ

˘ œ˘ œm

˘

5 5 5 5 5

œb>

f

œ

œm.

œn. œ

.œm>

œ œb.

œ. œµ

.œm>

œ œb.

œ. œ

.

œm>

œ œb.

œ. œ

.

œm>

œ œb.

œ. œ

.

55 5 5 5

œb

f

œ

œb

˚

œm

œ

œm

f

œm

œ

œn œ œm œm

œ

œn œ œm œm

œ

œn œ œm œm

œ

œn œ œm œm

œ

œn œ

5 5 5 5 5

œB˚

f

œ˚ œn

˚œb˚

œµ˚

∑ ÷

/ 156 / 160/ 152

Ó

œœœœ

œœ

m

µµµ

œ

œœ

œœ

Â

µ

µ

œœœœœ

œ

œœ

µ

Bm

µn

œµœ

œmœµ

ϵ

“big pluck”

œm

f

œœm

œmœ

œµ˘

f

œb˘

œ˘

œµ˘ œ

˘

œµ˘

œb˘

œ˘

œµ˘ œ

˘

œµ˘

œb˘

œ˘

œµ˘ œ

˘

œµ˘

œb˘

œ˘

œµ˘ œ

˘

œµ˘

œb˘

œ˘

œµ˘ œ

˘

5 5 5 5 5

œ

œ

˚

f

Œ

œ˚ œ

˚œ˚

&

Œ œ

œ

b

f

œ

œm

œœm

œ

j

œ

œb-

f

j

œ

œb-

j

œ

œb-

j

œ

œb-

j

œ

œb-

.˙m

Ï

Ï

.˙m

Ï

Ï

.˙bB

Ï

.˙bB

Ï

.˙B

Ï

p

sus. cym.

/ 164

noise wave

œ

.˙>

.

.

˙

˙

æ

Ï

.

.

˙

˙

Ï

.

.

˙

˙b

æ

Ï

.˙m

Ï

23Tear of the Clouds

3/16/08

A. Tear of the Clouds score 131

Page 142: Klingbeil Dissertation Web

&

&

&

B

&

&

?

÷

&

15

&

?

&

&

15

&

?

&

&

&

?

4

5

4

5

4

5

4

5

4

2

4

2

4

2

4

2

4

4

4

4

4

4

4

4

4

3

4

3

4

3

4

3

8

5

8

5

8

5

8

5

4

4

4

4

4

4

4

4

Fl.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~

93

w

Œ

take Piccolo

93 .˙ œœµ

f

œœm

œ

œ œm

6

93˙ œ œ œbB

f

œ œ œb

œ

œb œ

œ

6

93.˙ œ œ

f

œœ

œ œm

5

?

93

w

J

œ‰

?

93.˙ œ œ œbB

>œ>

œ>

œ>

œ>

R

œ

3 3

93 w

J

œ

93

w

J

œ

f

93

93

w

Œ

93

.

.

˙

˙

æ

.

.

œ

œ

æ

œ

f

œœb œ œb

œB œ

6

.

.

˙

˙

æ.

.

œ

œ

ϵ

f

œ œ œ œœ

5

B

w

w

æ

œ

f

œnœ œµ

œ

5

w

ϵ

f

œ œœ

œ

>

Ï

œ>

œb

>

Ï

œm>

œ

>

Ï

œ>

œ

>œbB>

ƒ

&

œb

>

œbB>

f

œ

>

œ>

ƒ

œ

>

Ï

œ>

f

Chinese cym.low conga

/ 168

œb

œ

“deep synth”

œb

œ

œm

˚

Ï

œ>

œ

Ï

œ

œ

B>

œb

Ï

œ

œb

n

>

œb

Ï

œm>

.˙ œ

⋲œm

œ

œm

3

.˙ œ

œ

œ

œ

3

.˙ œ

⋲ œ

œ œm

5

œm

flœ

fl

œ

fl

⋲?

.˙ œ

œ

˘

œ

˘

.˙ œ

Ó Œ‰ œ

>

f

œ>

œ>

œ>

6

hi bongo

Œ

Œ

˙ œ œ

f

œœm œB œ

3

˙

˙

œ

œ

œ

œ

f

⋲œ œ

œ

5

˙

˙

œ

œ

œ

œ

f

œ œµœ

5

.˙ œ

f

œ œœ

˙ œ

.œ>

ƒ

˙b œ

.œm>

ƒ

˙ œ

>

ƒ

˙

Ï

œ.œbB

>

ƒ

&

˙b

Ï

œ

.œb

>

ƒ

˙

Ï

œ

.œ>

ƒ

˙æ

>

.

J

œ

F

˙b œ

˙b œ

˙m

˚

œ

.œ>

ƒ

˙

œ

.

.

œ

œ

B>

ƒ

˙b

œ .

.

œ

œb

n

>

ƒ

˙b

œ

.œm>

ƒ

.œ œ

.œ œ

.œ œ

.œ œ ?

.œ œ

.œ œ

.œ œ

.œ œ

.œ œ

.

.

œ

œ

œ

œ

.

.

œ

œ

œ

œ

.œ œ

24 Tear of the Clouds

3/16/08

A. Tear of the Clouds score 132

Page 143: Klingbeil Dissertation Web

&

&

&

?

?

&

?

÷

&

15

&

?

&

&

15

&

?

&

B

&

?

4

4

4

4

4

4

4

4

8

5

8

5

8

5

8

5

8

9

8

9

8

9

8

9

4

4

4

4

4

4

4

4

Fl.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

98

98

œ

>

œ w

98

œ

wb

98

œ

>

œm

w

98

flz.

f

&

98

wb

æ

flz.

f

98

w

æ

flz.

f

98

.˙æ

f

Œ

98

wb

98

wb

98œb œ

wm

œmœ

w

œµœ

wb

œ œwb

œ5

F

j

œ

œ

ƒ

œm˚

F

j

œ

œb

ƒ

œ

F

j

œ

œ

ƒ

œbB˚

F

œ

ƒ

.˙?

&

œb

F

œb

ƒ

œ˚

F

œ

ƒ

œ

fl

ƒ

œ

f

˙Œ

low bongo

œ

œb .˙

œ

œb .˙

œ˚ j

œ

œm .˙

œ

œ

j

œ

œ .˙

œ

œb

j

œµœb .˙

œµ˚

j

œ

œb .˙

Ï

œ

.œm

Ï

œ

Ï

œ

.ϵ

æ

Ï

flz.œ

æ

.œb

æ

ƒ

flz.œ

æ

æ

Ï

flz.œ

æ

.x

æ

f

x

æ

temple blocks

(senza dim.)

.œ œ

.œ œ

æ

œ

æ

.

.

œ

œ

B

æ

œ

œ

æ

.

.

œ

œb

æœ

œ

æ

œµ œ

p

J

œm-

P

œm.

F

œ

>

œ

>

œ œµœ œ

œ œmœ

ƒ

J

œB

6

œ-

P

œm.

F

œb

>

œn

>

œœµ

œb œ

œmœ

œ

ƒ

J

œB

6

œ-

P

œm.

F

œm> œ

œm œm œœ œm œm

ƒ

J

œ

6

œm-

p

œ.

.œb

>

œ œ

F

j

œ

œm-

p

œ.

>

œ œ

F

J

œ

œ-

p

œb. .œµ

>œ œ

F

J

œ

j

œ-

p

‰ Ó Œ .

c1 / 12

.

.

˙

˙

Π.

.

.

˙

˙

Π.

ϵ-

P

œ

FF

œm- œ

-

œ œµ œ œm œœ

ϵ

ƒ

J

œ

6

œb-

P

ϵϵ-

ϵ-

œœ œm

œµ œÂœ

œÂ

ƒ

J

œ

6

&

œ-

P

œb

F

ϵ-

ϵ-

œ œÂ œµ œ œ œm œ

ƒ

j

œn

6

B

˙µ

f

Œ

p

Π.

25Tear of the Clouds

3/16/08

A. Tear of the Clouds score 133

Page 144: Klingbeil Dissertation Web

&

&

&

?

&

&

?

÷

&

15

&

?

&

&

15

&

?

&

&

B

?

4

4

4

4

4

4

4

4

4

2

4

2

4

2

4

2

4

4

4

4

4

4

4

4

Picc.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

102

Œ ‰j

œ

p

œœ œ œ œ

œ

f

œm œœm œ

œ œm

p

6

6

Piccolo

102œ

œ œœ

œ

P

œm

>

f

œœ œ œ

œœ œB

ƒ

œ œœ œ œB

56

6

102œB

œœb

œ

œ

P

œm

>

f

œ

œ

œB

œ œ

œn

ƒ

œBœ œ

œbœn

œ

5 6 6

102 œbœ œ œ œm

P

œm>

f

œ œ œb œ œb

œbœn œn

œnœn

œm55

6

102

j

œ

p

‰ ‰

j

œB

p

œ œ

F

102

J

œ

p

‰ ‰

j

œb

p

œ œ

F

102

J

œ

p

.˙µ

p

102

102c3 / 20

Ó

˙˙˙

˙˙˙

µ

µB

m

Ó˙m

102

102 œœbB

œmœ

ϵ

F

ϵ

f

œ

æ

œ

æ

œ

æ

œÂ

æ

œœµ

œµ œBœ œµ

5

6

œ@

œ@

ϵ

@

œ

@

ϵ

f

œm

æ

œ

æ

œm

æ

œBœm

œµœ

œm

3 5

B

ϵ

@

œb

@

œ

@ œ@

œ

f

œ

œm

æ

œB

æ

œœm

œµœ œb

œ

56

?

Œ

˙

f

œœ

œµ œÂœ

5

q = 108

q = 108

r

œ

‰ . ‰œm œ œ œm œ Œ

6

œ>

f

œ œBœ œ

Œ ‰ ‰

œ

œœm

œµ œ

6

3

3

œb

>

f

œBœb

œb œ‰

œ

œmœn

ϵ

J

œ

œœ œ

ϵ

J

œ

6 6

3

3

œ

ƒ

Œ Ó

j

œB

fl

S

‰ Œ Ó

J

œb

˘

S

‰ Œ Ó

J

œ

˘

S

‰ Œ Ó

J

œ˘

S

‰ Œ Ó

hi bongo

œ

œœœœ

œ

b

B

µ

b

B‰

œœœœœœ

œ

n

n

m

m

µµ

n

Œ

j

œ

œœœœœœ

µ

nµBB Œ

3 3

œ

œ

œ

Â

b

‰œ

Œ

J

œœœ

œ

Œ

33

“big pluck”

œ

œœb

œœ

œ

bb

F

œœœmm

œœ

œ

j

œ

œœ

nm

J

œœœ

nm

Œ

3 3

œ> œÂ

œB œµ

‰ ‰

œ

œ

œµœ œ

r

œ

˝pizz.

Œ

6

6

3

œm

>

œœµ

œB

‰ ‰

œ

˝

pizz.

‰ ‰

r

œ

arco

œ œ

œb œ

6

33

3

œÂ

‰ œ

œ

œBœm œ œ

œ

œ œµœm œ

6 3

3

œœ ‰

œm

˝

pizz.

‰ ‰

r

œb

˝

Œ

3 3

‰ œ

˚

ƒ

œb

˚œµ˚ œ

˚ œm˚

œ

˚

ƒ

œb

˚ œÂ˚ œµ

˚ œ˚

‰ œ

˚

ƒ

œm˚ œ

˚ œ˚

œm˚

œm

˚

ƒ

œb

˚

œm˚

œ˚ œbB

˚

/ 28

.

.

.

.

.

.

.

œ

œœ

œœ

œœ

µ

b

Â

B

.

.

œ

œ

‰ J

œ

œœ

b

Œ

J

œœ

œmŒ

‰œÂ

´

arco

ƒ

œ

´œ´

œb´ œµ´

j

œ

œ

˝pizz.

Œ

j

œÂ

˝pizz.

Œ

j

ϵ

˝

Œ

26 Tear of the Clouds

3/16/08

A. Tear of the Clouds score 134

Page 145: Klingbeil Dissertation Web

&

&

&

?

&

&

?

÷

&

15

&

?

&

&

15

&

?

&

B

?

?

4

4

4

4

4

4

4

4

4

2

4

2

4

2

4

2

4

5

4

5

4

5

4

5

4

4

4

4

4

4

4

4

Picc.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

105

⋲œ

œmœ œm

⋲œb œn

œ œn œm

⋲œ

œbœ

œ œµœb

J

œ

5 6 7

105

⋲ œmœm œ

œm

⋲œ

œmœ

œbœµ

œ

œbœ

œb œnœn

J

œm

5 6 7

105

Œ ⋲ œb

œb

œB

⋲ œn

œbœ

œB

R

œ

⋲j

œ

S

5

105

⋲ œ

œmœm

œ

Œ ⋲

œm

œœn

œm

‰j

œn

S

55

105

Œ ⋲.

j

œm

^

S

Œ ‰j

œ

^

S

105

Ó Œ ‰ J

œbB^

S

105

Œ ⋲

.

j

ϵ

^

S

Œ ‰ J

ϵ^

S

105

∑&

105

œœœœ

B

m

⋲ ....

j

œœœ

œ

Ó

/ 30/ 34 / 36

œœ

œœœµ

m

B

.

.

.

.

J

œ

œœœB

Bb

m Œ

œœœ

œœ

œ

œ

œ

µ

B

b

µ

µ

µ

œ

œ

œ

œ

œ

œ

œbB

Â

µ

n

105

œ

œm

⋲ ...

j

œœœ

Œœœ

œ

œ

œœ

œ

m

.

.

.

J

œ

œ

œ

bŒ œ

œm

œ

œm

ƒ

105

Œ ⋲

œÂ

œ œµœµ œµ

œ

œB

œ œµœ

œ

6 5

⋲ œÂ

arcoϵ

œœm œ

œ

œÂ

œ

œµœ

œn

55

⋲œµ

arcoϵ

œÂ œÂ

⋲ œ

œœÂ

J

œ

œm œm

5

Œ ⋲

.

J

ϵ

pizz.

Œ œm

œ

Πϵ

f

œ œm

3

ϵ

f

œ

ϵϵ

œmœ

3

3

Œ

œ

f

‰ j

œb

p

œ

œ#

vibraphone

(senza pedale)

œœ

œœm

µ

œ

œm

œ

ƒ

œ

œ œm œœµ œm

7 7

‰ r

œ

ƒ

œ

œ

œ œ

8:5œ

r

&

œ

.œœ œ

œmœm

5

5

Ó

⋲ .œ

f

œ œ œœb œ

œœ œm

œ œ œœ

œmœ

œm œnœ œb œn œ œn œm

œ œ œm œ œm

ƒ5

6

7

œµ œ œmœ œ

œÂ

f

œÂ œmœ œm œm

œµ œn œb œ œ œm

œ œµ œ œb œb œn œ œm œ

Ï5 6

7

R

ϵ

‰ .

ϵ

f

œœ œµ œm

œ

œn œ œ œm œœ œµ œ

œm œn œb œn œœµ œ

Ï

5

67

œ

>

ƒ

œ œ

œ

>

œ œ

>

œ

>œm œ

œb œmœ

Ï

7

wbB

p

œ

ƒ

Œ

p

œ

ƒ

w

p

œ

f

œ#

œœb œ#

œn œœ#

œœ

œbœ#

œn œ œb ⋲⋲ œn

œ œœ3 5

/ 44

œmœ œµ

œ œ œµ œm œ œ

œµ œœ

œmœn

œ œµœµ

œœB œ œ œm œn

œm œµœ

œ œ

5 5 6 67

œµ œm œµ œ œ œµ œ œmœ

œmœm

œœ œµ œm

œ œmœ

œ œ œb œ œm œ œ5 6

7

œœ

>

ƒ

œ

œœ

œb œm œœµ œ

œ

œœb œn œ œ œm

œµ œ œmœ œm œ

7 55

6

6

&

ww

ƒ

arco

œœ

27Tear of the Clouds

3/16/08

A. Tear of the Clouds score 135

Page 146: Klingbeil Dissertation Web

&

&

&

?

&

&

?

&

&

15

&

?

&

&

15

&

?

&

&

&

?

4

4

4

4

4

4

4

4

4

2

4

2

4

2

4

2

4

4

4

4

4

4

4

4

Picc.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

108

Œ ‰

œm

f

œ œm œm œœm œm œm

œbœn œ œn œm œ œ œb

3

6

7

108

ƒ

œmœB

œœm œm œm

œ

œm œm œmœ œ

œ

œn œm œmœm œ œ œ .œ

56

67

108

ƒ

œ

œmœB

œœµ œm

œnœ œ œm

œ œ

œmœ œµ œm

œ œmœb œ .œ5

6 6

7

108

Ó ⋲

œ

flƒ

œm

fl

œ

fl

œm

fl

œœ

œ œ œ5

6

B

108

Ó ‰

J

œ

>

Í

œ

108

Ó ‰ J

œ>

Í

œ

108

Ó ⋲ r

œ

f

gliss.

œ

ϵ

œœµ

œmœ

œ œµœ

ƒ

3

6

VII I

108

œ

f

œ#

œœ# œ#

œœn œ#

J

œ‰

œ

°

œœ

œ#

œ œ#

53

6

108

Ó

œBœ

œÂ

œ œBœ

œ

7

/ 50/ 48/ 46 + / 52

Ó ‰œ œ

œ œm

6

œm œ

œm œ œ œ

108

Ó

œœ

œ

œ œœ

œ

7

Ó ‰œ œ

œ œm

6

œm

f

‰ j

œm

Ó

108⋲

œœµ œm œÂ

œ œ œœÂ

œ œ œ œµœm

sul Eœ

æ

æ

6 6

œm œœm

œ

œBœm

œ

ϵ Ͼ

Ͼ

sul C—

æ

5

B

⋲œ

œBœm

œ

œœ

œB

œœ

.œÂ

æ

sul G

œ

æ

—Â

æ

5

?

Œ œm

æ

œm

æ

œ

æ

sul G

æ

æ

˙m

ƒ

˙˚

Ï

˙˚

Ï

˙˚

Ï

˙

ƒ

˙

ƒ

˙

˙>

ƒ

˙

Ó

˙

˙

Ï

˙

˙

B

Ï

Ó

wµ˚

wm˚

wm˚

wm˚

wb

˚

Ï

wB˚

Ï

œ

œ#

* °

Œ Ó

/ 54

wb

w

w

m

w

w

µ

&

∑&

˙b

f

˙

œb>

f

œ œm œm

œ

R

œm>

ƒ

‰ . Ó

5

œ>

Ï

œÂ œµœµ

œ œ ˙

ƒ

5

œ>

Ï

œmœ

œBœ

œbB ˙

ƒ

5

œ>

ƒ

œ

œm œœm

œ ˙

F5

Œ.˙

p

Œ

.˙b

p

Ó

˙b

p

œ

œ

œ

œ œBœ

œÂ œm

5

/ 56

œbœ œ

œBœm

œœ

µm

Œ

œœ

œœ

œ

n

Â

n

µB

5

Ó Œ

œœ

œ

µb

œ

œ œœ

œ œmÓ

5“bells”

œbœ œ

œ

œmœm

Ó

5

Ó .

œ

œ

ƒ

œÂ œµœµ

œ œ ˙

F

5

œm

ƒ

œ

œ

œmœm

œ ˙

F5

œ

ƒ

œ œb œœ œµ ˙

F5

œm

˙

ƒ

œ

sul pont.

q = 66

q = 66

p

w

w

w

w

ƒ

w

ƒ

w

Ï

brassmallets

w

w

w

w

28 Tear of the Clouds

3/16/08

A. Tear of the Clouds score 136

Page 147: Klingbeil Dissertation Web

&

&

&

B

&

&

?

&

&

15

&

?

&

&

15

&

?

&

&

&

?

4

3

4

3

4

3

4

3

4

2

4

2

4

2

4

2

Picc.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

D

D

113wb

Ï

113 œ

Œ Ó

113 ϵ

Œ Ó

113œ

Œ Ó

113

œŒ Ó

113

œ

Œ Ó

113

œ

Œ Ó

113 œœbb

ƒ

Œ Ó

crotales

113 œœœœœœ

b

B

bµb

Œ Ó

113

œœœœœœ

mH

ƒ

Œ Ó

113 O

w

f

O

w

µµ

f

œ

Œ Ó

œ

ord.

Œ Ó

ø

ø

w

Ó

˙b

ø

O

w

O

w

w

Ï

29Tear of the Clouds

3/16/08

A. Tear of the Clouds score 137

Page 148: Klingbeil Dissertation Web

&

&

&

B

&

&

?

&

&

15

&

?

&

&

15

&

?

&

&

&

?

4

2

4

2

4

2

4

2

4

4

4

4

4

4

4

4

4

3

4

3

4

3

4

3

4

4

4

4

4

4

4

4

Picc.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

Ÿ~~~~~~~~~~

Ÿ~~~~~~~~~~~~~119

Œ ‰ . R

œb>

F

119

119

119

119

119

119

119

119

œ

œœœ

œœœ

µ

m

µ

B

p

Œ

119

œœ

œ

“bowed pluck”

œœœ

m

p

Œ

119

J

œm

pizz.

p

‰ Œ

j

œB

pizz.

p

‰ Œ

œb( ).˙ œ˘

œm

œ

5

œ

œ

b

p

Ó .

crotales

œœœœœ

œB

B

bBB

Œ Ó

œœœœœ

b

œ

Œ Ó

‰J

œB

ø

arco.˙

j

O

ϵ

ø

arco

.

.O

˙

j

O

œ

mm

ø

.

.O

˙

π

.

.O

˙

p

.

.O

˙

p

œ>

.œb

ƒ

.

.

J

œœbb

f

Ó .

.˙ .œ

œ

F

softmallets

∑?

œm( )œ œ

œbœ œm

œm

œ ˙

p

6

œ

œ

œ

#

F

°

Œ ‰

œ

Œ

3

vibraphonemotor on (slow)

œœ

œœœ

œ

b

B

p

œœB

m

œ

Ó

œ

œœ

b

p

Œ Ó

œœm

œ

Ó

J

œ

œ

µ

pizz.

P

‰ Œ ŒO

œ

BB

ø

arco

J

œb

pizz.

P

‰ Œ Œ O

œ

nm

ø

arco

J

œ

œ

B

pizz.

p

‰ Œ Œ

O

œ

µµ

ø

arco

Œ œ

P

pizz.

Œœ

osul G

ø

arco

&

30 Tear of the Clouds

3/16/08

A. Tear of the Clouds score 138

Page 149: Klingbeil Dissertation Web

&

&

&

B

&

&

?

&

&

15

&

?

&

&

15

&

?

&

&

?

&

4

3

4

3

4

3

4

3

Picc.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

Ÿ~~~~~~~~~~~~~~~~~

p

p

p

p

125

.œ œ

œm>

f

œm œœm

œ

œ œ œ

.œb5

7

125

125

125

125

125

125

125

Œ ‰ œ ⋲

.

.

J

œ

œ# Œ

3

125

Œ ‰

j

œœœœœœœœ

œ

µ

m

µ

m

b

µ

p

Ó ⋲

.

j

ϵ

Œ

125

Œ ‰j

œœœœ

œ

mmm

Ó

Œ ‰

J

œ

œ

b Ó

Ó ⋲

.

j

œm

Œ

125

O

w

O

w

O

w

w

o

˙

œ

p

‰ .

R

œ>

.œ#

°*

p

œœ

œ

J

œ

5

œœœ

œB

µ

b

œœ

œ

µµ

π

œB

p

Œ Œ ‰ j

œb

œœ

œ

mm

Œ Ó

œ

Œ Œ ‰ j

œb

F

œ( ).˙

F

.œb

fl

f

œbR

œ5:3

œœ#

œb œ

œbœb

3

3

.˙ œ

p

⋲œ˘

°*

œb

œ

œœ

œb

7

œœœœœ

µ

µµ

œœœœœ

bb

œœœœm

Œ Ó

œœœœœ

bb

p

Œ Ó

J

œ

œ

m

pizz.

p

‰ Œ Ó

J

ϵ

pizz.

p

‰ Œ Ó

J

œ

pizz.

p

‰ Œ Ó

˙ .

J

œ

œ œ œbœ œ œm

Œ

œ

°*

œœ

œb

5

Œ

œœœµ

Œ

œ

œ

œœ

œ

m

µ

b

m

Œ ⋲ .

j

œB

p

Ó

Œœ

œœ

m

Œ

œb

.œÓ

ΠJ

ϵ

p

‰ Ó

31Tear of the Clouds

3/16/08

A. Tear of the Clouds score 139

Page 150: Klingbeil Dissertation Web

&

&

&

B

&

&

?

&

&

15

&

?

&

&

15

&

?

&

&

?

&

4

3

4

3

4

3

4

3

4

2

4

2

4

2

4

2

4

4

4

4

4

4

4

4

4

2

4

2

4

2

4

2

Picc.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

Ÿ~~~~~~~~~~~~~~~~130

œm( )˙

F

.œœ

130

130

130

130

130

130

130œb œ œ

œ

œ

3

130

130

130

∑?

œ>

f

œ

œmœ

œ œb œ˙m

7

œb œœ#

œ œbœ

Ó

œ

œ

œ

œ

µ

b

B

œ Œ

œ

œœœœ

œ

m

m

µ

B

πœ

œœœ

œ

ϵ

Bb

p

Ó œ

œ

m

œ Œœœœ

œ

m

œ

œœb

Œ Œ

Ó

J

œ

œ

µ

m

π

j

œ

p

‰ Œ

J

œB

π

J

œ

œ

µ

p

‰ Ó

j

ϵ

p

pizz.

‰ Ó

œ . .œ

p

œm

F

œ

œœ

œœ

‰3

3

.œ œn œb

œ œ

.œb œ œ

J

œ

7 5 3

œ#

°*

F

œ œ#

œ œ

Œ

˙

˙n

p

œB

c11 / 60

œœœœœ

µb

µµ

‰ .

r

œœœ

œ

œœ

œ

B

B

B

b

µ

œB

p

œœœ

œ

b

bb

Œœœ

œ

œ

Œ Ó

˙Â

π

˙Â

p

˙µ

π

˙

p

O

˙

bb

π

O

˙

BB

p

œm>

f

œœ œm œ œ

œb œmœ

œm œœn œ

7

3

œ

œ œ j

œ

7

Œ

œm

˙

˙

O

˙

Œœm

P

32 Tear of the Clouds

3/16/08

A. Tear of the Clouds score 140

Page 151: Klingbeil Dissertation Web

&

&

&

B

&

&

?

&

&

15

&

?

&

&

15

&

?

&

&

?

?

4

4

4

4

4

4

4

4

4

3

4

3

4

3

4

3

Picc.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

135œm œ œ œ

œœm

œB œ

135

135

135

135

135

135

135

œb

œ œb

7

135 c12 / 66

.

.

.

.

.

J

œ

œ

œ

œ

œb

b

µ

b

135

.

.

j

œ

œ

m

b

Œ

⋲.

.

J

œ

œ

135œ

ϵ

œœµ

O

œ

O

œ

w

p

œ

œ œ

œ œb

5

c13 / 72

Œ

œ

œœ

œ

œ

B

B

Œ

œ

œœ

œ

œ

Ó

Œ

ø

Œ

ø

.

.

O

˙ Œ

ø

w

˙

˙#

Œ œ

3

c15 / 84 c16 / 90

Ó

œ

œœ

µ

Ó

ϵ

œ

π

Ó

œm

œ

Ó Œ O

œ

µµ

ø

Ó

O

˙

ø

Ó Œ

œ

p

Œ

ø

Œ ‰

˙

j

O

O

m

ø

sul Asul D

.

.

O

O

.

.O

˙

O

œ

p

O

˙

O

˙

p

ø

O

O

p

.

.O

˙Œ

ø

O

˙

Ó

ø

33Tear of the Clouds

3/16/08

A. Tear of the Clouds score 141

Page 152: Klingbeil Dissertation Web

&

&

&

B

&

&

?

&

&

15

&

?

&

&

15

&

?

&

&

?

?

4

3

4

3

4

3

4

3

4

4

4

4

4

4

4

4

Picc.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

q = 90

q = 90

141

Ӝ

Ï

œb œ œm œœm

œ œm

141

141

141

141

141

141

141

Ó

œ

°

f

*

œœ# œ

œ

hard mallets

141

Ó ®

œm

f

œµœµ œµ œm œµ œ

141

141

q = 60

q = 60

E

E

œ>

f

π

œ>

Œ Ó

˙

π

˙B

c1

˙

π

°

˙b"slow synth”

˙ ˙m

˙m˙ ˙µ

3

˙˙ ˙m

3∑

‰ J

O

œ

ø

.

.

O

˙

.˙ œµ

œ œb œ œµ œµ

5

œ œb œ œm œm

5

.

.

O

˙

O

œ

J

O

œm

p

3

.˙ .œ œb

œ œµœb

Ó

3

Ó

œm œm œµ œb

œ œmœb

Ó

3

Ó

œm œm œn œb

.

.O

˙

O

œ

O

œ

.˙ .œ œ

O

œ

O

˙

O

œ O

œ

mm

J

O

œ

3

∑ B

34 Tear of the Clouds

3/16/08

A. Tear of the Clouds score 142

Page 153: Klingbeil Dissertation Web

&

&

&

B

&

&

?

&

&

15

&

?

&

&

15

&

?

&

B

?

?

Picc.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

*

147 .˙ œ .œµ

5

147

147

147

147

147

147

147

take bow

147

Ó

˙B

π

147

Ó

˙b

π

°*

“bowed synth pluck”

147O

˙

O

œ

p

.

.

O

œ

µµ

O

œ

Œ Œ

J

O

œ

ø

O

˙

3

˙ .œ œ œ

5

˙µ˙ ˙µ

3

˙˙ ˙m

3∑

O

œ ..

O

œ

bb

O

˙

O

œ

O

˙ . O

œ

5

œ

J

œ œb .œ œ œ

3 3

œ œb œ œµ

œ œb œ œm

O

sul A

P

J

O..

O

ϵ

.

.O

˙

.

.O

œO

œ

µµ

p

˙ œµ

p

. .œ œÂ

œµ œœÂ œn œb

5

œm œ œ œ œb

5

O

œ

r

O

œ

µÂ

O

˙

O

œ

5

.

.

O

˙

.

.

O

œO

œ

5

œ .œ œn œ œ

5

Ó Œ ‰J

œ

Ó Œ ‰ J

œ

π

crotalearco

Œ

œB œµœ

œm œm œµ œbœµ œn

6

Œ

œb œœ

œm œm œn œbœm œn

6

O

œ

O

sul A

.O

ϵ

p

sul tasto.œ

O

˙. .. .

O

œO

œ

bb

O

œ

35Tear of the Clouds

3/16/08

A. Tear of the Clouds score 143

Page 154: Klingbeil Dissertation Web

&

&

&

B

&

&

?

&

&

15

&

?

&

&

15

&

?

&

B

?

?

Picc.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

152

J

œ œµ œ œÂ

J

œ œn

J

œb

3

152 .˙ œ.œµ

152

152

152

152

152

152œ

Œ Ó

152

œµ œ œb œ

π( )œµ

5

œB œµ œBœ

Ó

Ó

ϵ

p

œn

œ

ϵ

152

œm œ œb œ

π( )œm

5

œ œ œœ

Ó

Ó

œm

p

œn

œ

œm

152 œ.œb œ

œµ œ .œm

5 5

.

.

O

œ

O

œO

œ

bb

O

˙

3

œ.œµ .œm œm œµ œb

‰ Œ

5 3

˙ œ œ œ

œœÂ œn œb

œm œm

6

œ

F

Œ Ó

œ œ œ œbœm œm

6

œ

F

Œ Ó

.œ œµ . .œ œm

P

sul pont.

œ œÂ

O

œ

ϵ

p

sul tasto.œ .œ œB œ

&

w

π

œ œ.œ œ .œ œµ

p

˙ ˙

c2

œµ œbœµ œn œB œµ œB

7

˙

π

˙

œ

°*

œbœm œn œ œm œ

7

œn œ

r

œ .œ œµ

j

œ œ

5 5 3

.œœµ œ

J

œ .œ

œ œ .œ ˙ .œ œµ

œ . .œ œµ œ œ

R

ϵ

5

˙µ ˙µ ˙µ

3

œ œBœB

œb œB œ œB

œm

˙m ˙n˙m

π( )

3

œ œœ

œb œ œ œ

œm

p

œB

j

œB

sul tasto.

œ .œb

‰ J

O

œ

ord.

ø

O

œ3

œ.œ

P

œ

sul pont.

œµ .œ œB . .œ

œ

B

˙

J

œ.œ

œ œ .œn

J

œ œœ œ

5 3

œbB

ø

˙ œ œb

3

.

J

œ

π

œ

Ó

Ó

˙

œBœµ

œmœÂ œ

5

Ó

˙

π

œ œ

œmœ

œm

5

.

.O

˙ O

œm

p

.œ œB œ

j

œB œœb œB

sul tasto.

œ œB

3

5

36 Tear of the Clouds

3/16/08

A. Tear of the Clouds score 144

Page 155: Klingbeil Dissertation Web

&

&

&

B

&

&

?

&

&

15

&

?

&

&

15

&

?

&

B

?

?

4

3

4

3

4

3

4

3

8

7

8

7

8

7

8

7

Picc.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

157œ œ œµ œ œ

R

œm

p

œ

3 5

157

œm

p

œ . .œn œ

r

œ .œ œµ

5 5

157 ˙ œ.œ

p

œ

157

157

157

157

157

157

˙ ˙µ ˙µ3

œB œ œ œ œ

5

157

˙ ˙m ˙n3

œ œ œ œ œ

5

157..

O

˙

O

œ.O

‰ J

O

œ

µµ

ord.

ø

O

˙

.

.

O

œO

œ

5

J

œ .œÂ œµ

J

œ œµ

3

œ œB œ œ œœ

π

.

J

œbB

π

œ

5

.œ œB œ œ œbB œ

5

œBœµ

œmœÂ

œµ œB œ œBœB

œb œB7

œ œ

œmœ

œm œn œ œœ

œb œ7

.O O

œ

O

˙ O

œ

mm

p

O

œ

œ.œm œ .œ œ

.œµ .œµ œ

5 55

œ œ.œb œ .œ

œ

P

œ.œ œ

J

œB œ œ

p

3

Ó Œ

œ

ø

con sord.

Ó Œ

ϵ

ø

con sord.

œ œB œ œ œ

5

œ œB

œm

Œ Ó

3

œ œ

π( )

œ œ œ

5

œ œ

œm

p

Œ Ó

3

O

˙

O

œ.

.

O

œ

µµ

O

œ

O

œ

O

œ . .. .

O

œ

µµ

O

œ

O

œO

œ

µµ

∑B

Œ

.˙m

ø

arco

œb œ

‰ ‰J

œ

˙

3

˙ œµ .œœµ

œb .œ œn œ.œ œB œ œB

53

.˙ œ

p

.˙ œ

p

p

œ

Ó

5

Œ

˙ œ

œ œµ œB œ œBœB6

Œ

˙

π

œ

œ œm œn œ œœ6

O

œ

R

O

œO

œ

O

œ

O

œ

R

O

œ

µµ

5 5

O

œ

O

œ

J

O

œ

mm

O

œ..

O

œO

œ

P

3 5

.

J

œB

ø

sul tasto.

˙ œ

5

.˙ œ

p

˙ œ œµ œ

3

œ œ

P

.œn .œœ

5

œBœb œ œ

r

œ

‰ .

œbB

ø

Œ ˙µ œµ

3

œb œ œ œ

œm

Œ ˙m

π( )

œn

3∑

œb œ œ œ

œm

P

O

˙

O

œO

œ

mm

P

j

O

œ

3

O

œ

J

O

œ O

œ

ÂÂ

O

œ

3

π

37Tear of the Clouds

3/16/08

A. Tear of the Clouds score 145

Page 156: Klingbeil Dissertation Web

&

&

&

B

&

&

?

&

&

15

&

?

&

&

15

&

?

&

B

B

?

8

7

8

7

8

7

8

7

4

4

4

4

4

4

4

4

Picc.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

162œ œ

.œm œ œ.œµ

p

162

œ œm œn

R

œ

J

œ

5

162 œ œ œb œ

J

œ

162

162

œ

Œ

ø

Π.

162

162œ

Œ

ø

Π.

162

162

‰œB

œµœm

162

‰œ œ

œm

162

O

œ

O

œ ..

O

œ

µµ

O

œ

j

O

œ

µµ

O

œ

r

O

œ

⋲J

œ

p

sul tasto..œ œm

J

œ

œ . .œB œ .œœB

J

œ

ø

.œœµ œ œB œµ

r

œ

5

œ

p

œ œÂ œn œ

œ

œm

π

.œ œµ3

œ.œB

P

œ

R

œ

J

œB

J

œ œ

R

œm

3 5

˙ ˙m ˙µ

3c3

œB œBœm œn œB

5

˙

π

˙m ˙n

3

œ

°*

œœm œn œ

5

O

œ

O

œ

œm

p

sul tasto.

œ

J

œ œ

3

œ œn œ

J

œ .œ œ œ œ

3 5

œ œ

J

œb œ œ

J

œB

P

sul pont.

3 3

?

.œ œm œ .œn œ

j

œb

‰ Œ

5 6take Flute

œ œ .œn .œ œµ

P

.œ œm

5

œ œ .œn œ œB œ

J

œb

P

3

∑?

.

J

œbB

π

˙

J

œœm

3

.

J

œ

p

œ

Ó

œÂ œn œm œµ

œ œ œbœm

œµœ

6

œ œ œm œn

œ œ œbœm œ

œ6

œÂ

R

œ œ œ.œÂ

P

sul pont.

œ œn œ

5 3

J

œ œµ

J

œÂ œn œB

sul pont.

œ

œm œÂ œn œB

π

3

6:5

&

œ œ .œ

J

œ œ œB œ

5 3

Ó Œ ‰

J

œ

ø

J

œ œ œn œ œB œbB

R

œB

3 5

.

J

œ

J

œ œ œ œ .œb œ œ œBœB

55

3

Ó Œœ

ø

˙ œ . .œn œ

Ó

˙µ

ø

Óœ œm

œµ œÂ œµ œµ œÂ

5

‰j

œµœ

Œ Ó

Ӝ

π

œm

œm œn

π( )

œœm œn

5

j

œbœ

P

Œ Ó

œ œb œ œ œB œµ œ

sul tasto.

œBœ

J

Om

sul Aord.

ø

5 5

O

œ

ø

ord.O

œ ..

O

œ

BB

O

œ

R

œ

œbB œ.œ œ .œm œµ

sul tasto.

œm œn

53

w

38 Tear of the Clouds

3/16/08

A. Tear of the Clouds score 146

Page 157: Klingbeil Dissertation Web

&

&

&

B

?

&

?

&

&

15

&

?

&

&

15

&

?

&

&

?

?

4

2

4

2

4

2

4

2

4

3

4

3

4

3

4

3

Fl.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

166

Ó ŒJ

ϵ

π

œ

3

Flute

166œ

J

œ

J

œ œm œn

P

œb œ œ œb œ œBœ œb

œ

p

3 3

166

œBœ

œ

œ

ø

œ .œbB

J

œ œb

3

166

∑B

166

˙ ˙

p

166œ œB œ œ œb

p

œ

166

˙ ˙

p

166

166

œµœÂ œn œm œµ

5

œm œn œB œ œ œbœÂ

7

166

œœ œ œm œn

5

œm œn œ œ œ œb œ

7

166

O O..

O

œ

BB

P

O

œ

J

O

œO

œ

µµ

p

O

œ

R

O

œO

œ

.

.O

œO

œ

bBB

3 5

O

œm

ord.

ø

O

œ

O

œ

.

.O

œ

bn

p

O

œ

3 5

w

p

.œ œ œ œ

J

œm œ

3

ŒJ

ϵ

π

œ

3

J

œœB

P

J

œ œB

R

œn œ .œm œn

3 3 5

Œ ⋲

.

J

œ

ø

œ œ œb

˙Ó

ø

œ .œ .œ œ œ œb

Œ

ø

(remove mute)

œµ œÂ œµ œµ œÂ œm

6

œ

œ

Œ Ó

Œ

ϵ

œµœ

‰ Œ

œm œn œœm œn

œm

6

œ

œ

Œ Ó

Œ

œm

œbœ

P

‰ Œ

.

.O

œO

œ

µµ

O

˙

œ

œm

p

sul tasto.

O

œ

O

œ..

O

œ

BB

O

œ O

œ

µµ

P

O

œ

j

O

œ

µµ

3

..

O

œ

O

œO

œ

P

O

œ..

O

œ

bb

5

w

J

œ œµ œ

R

œn œ œµ

p

3 5

œµ œ

. .œÂ . .œm

P

œn .œ œb

5

œ.œb œ œ œb œ

œB œB œb5

5

œ œb œ

p

œ œÂ

œ .œœ œ

j

œ œ

3

œ œB œ œ œbœÂ

œ

œ

œ œ œ œ œb œ

œ

œ

œ

œ

œ

œn

B .

.

œ

œ

œ

œ

bB œ

œ

O

œ

j

O

œ

BB

j

O

œ

® .œm

p

sul tasto.œ .œn

œ

J

œ

3

53

.

.O

œ

œ

p

sul tasto.œ

R

œµ .œ œn œœB

P

5

B

œÓ .

ø

R

œ œµ œœ œ

5 3

œ .œ œm œb œ œb œ

6

œBœ

œ œb

.

J

œ

ø

œ.œ .œ

œÂ

5 5

‰ J

œ

π

œ

œ œb

c4

œµœ œµ œm œµ

œ

6

œ

π

œb

œmœ œm œm œn

œ

P

6

œ

œ.

.

œ

œ

B œ

œ

J

œ

œ

3

œ œb œ œbœb

sul pont.

œ

π

5

B

R

œœm œµ

œµœ?

Œ œ

ø

39Tear of the Clouds

3/16/08

A. Tear of the Clouds score 147

Page 158: Klingbeil Dissertation Web

&

&

&

B

?

&

?

&

?

&

15

&

?

&

&

15

&

?

&

B

?

?

4

3

4

3

4

3

4

3

Fl.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

170.œ

p

œµ œ.œÂ œm œµ

5

170

œ œ œµ

p

‰ ⋲

œ

π

œ. .œµ

5

5

170 œ œbB

J

œœµ

p

3

170œ œµ

J

œ œm œµœµ

3

?

170

Ó œm

ø

170.œ œbB œ .œ œB

170

170

170

.œ.œµ .œ

.ϵ

170

.œ.œm .œ

.œm

170. .

. .

œ

œ

œ

œm

œ

œ

œ

œm

µ

F

sul pont.

O

œ

π

ord..

.

O

œ O

œ

O

œ

ϵ

sul pont.

œm œµ

œ

π

‰ . .RÔ

O

œ

µµ

ord.

ø

O

œ

˙ œ

p

œb œ œm œn œ œB œµ œm

π

‰J

œb

π

.œµ œB œ .œ œ

F

ϵ

5

œ

J

œB .œ œ .œ œ

F

3

œ œµ œm œn

œ œ

p

œb

π3

B

˙ œ

p

œ œ

œm

P

œ

.

J

œ

π

œ œ œœ

marimbasoft mallets

Œ œ œb

œB œ œœµ

ϵ

5:3

Œ œ

p

°*

œb

œ œ œ œ

œ

5:3

. .

. .

œ

œ

µ

œ

mn

.

œ

œ

œ

B œ

œ.

.

œ

œ

b

b

5

J

O

œ O

œ

µµ

.

.

O

œ O

œ

P

O

œ

3

O

œ . .O

p

O .

.

O

œ

n O

œ ..

O

œ

ÂÂ

œ œ

f

œ

ø

.œ œB œ œ

œÂ œÂ œm œB œb œ œb œ œ

p

‰ .R

œ

π

.œ œB

J

œ œBœm œb

5 3

œ œ

J

œ œ .œb

R

œ œ

3 5

œb

J

œ œ œ .œm

3

5

œ œ

J

œ œ.œ# œb

3

5

.œ.œµ .œ

.ϵ

œb œ œb œ œ œB

.œ.œm .œ

.œm

œb œ œb œ œ œ

œ œb œ œ œB œµ

sul tasto.

œm œµœµ

π

3

O

ϵ

p

sul tasto.œm

R

ϵ

5

R

O

œ

 O

œ

ÂÂ

P

.œb

sul tasto.

p

œ œ .œb5

5

B

. .œµ œ .œ œµ .œ œµ

5

œ œµ

J

œ œµ.œB œ

F

3 5

œ œb œ œ œB œµ œm œBœ

œ

π

⋲R

œ

ø

P

œB œ œµ œ .œm œµ

6

œ Ó

ø

(remove mute)

œ .œn œ œ œœ œ œm

5 6

∑B

œ.œ œb

F

œ .œ.œ œ

œ# œ

67

3

œ œ œœ

ϵ

5

œm œµœµ

œ œµ œm œµ

œ

6

3

œ œ œ œ

œ

5

œm œmœm

œ œm œm œn

œ

π

6

3

Œ

O

œ

ord.

ø

J

O

œ

J

O

œ

bb

p

.œ œb œ œ œ œb

F

sul pont.

œ œ œ

5 3 5

œ œ œB

F

sul pont.

œµ œm œµœµ

œ œµ œm3

5

?

40 Tear of the Clouds

3/16/08

A. Tear of the Clouds score 148

Page 159: Klingbeil Dissertation Web

&

&

&

B

?

&

B

&

?

&

15

&

?

&

&

15

&

?

&

B

?

?

Fl.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

174

J

œ œB

P

R

œ œ œ œµ œµ œB

73

174œ .œ œµ œ

.œB .œB œm œ œB œb œ œb

6

5

174 . .œ œbB

p

œ

R

œµ .œ œB

5

174

œ .œ.œµ

œ œµ œm

œn

œm œn œ

p7

35

? B

174

174

œœ

π

œbB

p

J

œ œB3

3

174œ

π

senza sord.˙

174

Œ ⋲

.

J

œ.œ

œ

œb

œ œb œb œ

π

Ó

5

174

œb œ œb œ œ œB

174

œb œ œb œ

P

œ œ

174O

œO

œ

BB

J

O

œO

œ

3

œ œB œµ œm œBœµ

œ

sul tasto.

œµ œm

π

.

J

œ

7

5

5&

œµ œ

sul tasto.

œbœ

π

‰ ⋲

.

J

œb

π

sul tasto.œ

.œ6B

œ œ œb œB œBœµ

œ

π

⋲J

œ

π

J

œœm

3

œ œ œ

p

p

œ

J

œ œb5

3

œ œb

F

œœµ œ œB œ .œ .œ

7

Œ ‰J

œb

π

.œ œB

œœ

P

œ œ œ .œb

5

œB

P

J

œ

3

œ.œ# œ#

œ# œn

F

œ œb

53

œœm œ œb œ

œ

c5

œœB

œ œµ œ

œ

Œ

6

œ

p

œm œ œb œœ

œ

°*

œœ œm œ

œ

π

Œ

6

O

œO

œm

.

.

O

œO

œ

P

.

.O

œ

O

œ

bb

5 5

J

œ œm œµ.œB .œµ œ

3

.œB œB

F

œ œ

œ œœB

œ œµ

7

?

Œ ˙

ø

œ œ œ œb œ .œœ

P

œ œb œb

6 3

œb œB

F

œµ œB œ œ œb œB œB œB

p

3

5

œb œB œB œB

œ

œm

π

.œb

ø

œ3

.œ œB œ

J

œB œ

œ

P

3

∑&

œ.œ œ œ œ

œ

π

.

J

œB

p

53

j

œn œB œ

3

?

œ œ œ

œœ

‰ .R

œ#

π

œœ# œn

5

Œ œ œbœ

p

‰ Œ

3

Ó œ

F( )œm

œb

F

œ œbœµ œB œ œ œb œB

63

Ó œ

p

œm

œb œ œbœm œ œ œ œb œ

63

O

œO

œ

j

O

œ

œ

p

sul tasto.

œ .œ œm3

R

œ œb œB œB œB

œ

œBœ œµ œm œn

53 5

B&

œ

sul pont.

r

œ

π

‰ .

.

.

O

œ

mm

p

O

œ

˙ œ

F ø

œB œµ œB œ œ œb œB œBœµ

œ

π

‰J

œ

p

Œ

œ

p

J

œ œ œb œb

3

.

J

œœ

p

œb œb œB

F

œµ œB

7 5

R

œœB

j

œ œ .œ.œ

œb

œ

p

7? B

Ó ‰

senza sord.

.

J

œ

p5

œ.œ

R

œ œ.œ

F

œb

5

œ .œ .œµ

J

œ œ

3

œœ œb

œ# œ#œ# œn œ œb

F

œ5

3

œ œb œœ

Œ

œBœµ

œ

œBœ

œb œ œb

5:3

3

œµ œ

œ

œ œb œœ

Œ

œ œ

œ

œœ

œb œ œb

5:3

3

œm œ

œ

œ œµ œB

F

œ œ œ œb œB œB

5

6

.

J

œm

p

ord.œB

J

œµ œ œ œb

F

sul pont.

3

O

œ O O

œ

nO

F

œB

sul tasto.œB

œn

3

41Tear of the Clouds

3/16/08

A. Tear of the Clouds score 149

Page 160: Klingbeil Dissertation Web

&

&

&

B

&

&

?

&

?

&

15

&

?

&

&

15

&

?

&

&

?

?

Fl.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

p

178œm œ œb œ

œ œb œb œB

5 3

178œ œB

f

œµ œB œ œ œb œB œB œB

p

5

3

178œ œ œ œB œ œB œB œB

œ

œm œ

π

⋲6

7

5

178 œb

p

œB œB

178

.œ œb œ œB

178

.œ.œ œ œ œ

œ

p

.

J

œB

p

75

178

œ

j

œ

p

‰ Œ J

ϵ

p

3

B

178

œ œ

œœ

œ œbœ

‰ .

œ#

p

œ# œnœ œb

5

178

Ó ‰ œœm

3

œµ œB œ œ œb œB œBŒ

7

Ó

œµœ

œBœ

178

Ó ‰ œ

p

œm

3

œm œ œ œ œb œ œŒ

7

Ó

œ

œœ

œ

178

œœµ

œœ

π

œ

p

ord.

J

œµ œ œ

7

3 3

œ.œB .œB

œB

œœB

œ

ord.

œµ œm œn

π

3

5B

&

œBœ œµ œ œ

j

r

œ

p

sul pont.

⋲ ⋲

.

.

J

O

œ

mm

p

ord.5

Œ ‰

J

œ

π

q = 72

q = 72

œm

f

œ œb œ œm œn œ œB œµ œ œb œ œb œ

π

56

‰J

œm

p

œ œb .œ .œµ œm

7

π

œm œ œb œ œ œµ œm œ

f

56

.œ œb

F

œ œ œm œn

œB.œB

œ œ

œ?

R

œœ .œb œ œ

.œm

5 5

œ œ

P

œ

œœ# œ œb œ

F

œb œ œ œbœ œb œ œb

œœ

Ϸ

œ œb œ œµ œ œb

Œ

6c6

Ó

œ œµ œµ œ

œ

Ó

œ œb œ œm œ œb

Œ

6

Ó

œ œm œm œ

œ

Ó

œ œb œ œµ œ

f

œb œ

3

.

J

ϵ

p

.œ .œb .œ œm

F

5

O

œ

O

œ

O

œ

œÂ

˙

f

œ

p

œ

P

œ œµ œ œb œœm

f

5

œ œ

f

œb œ œb œ œ œbœµ œ œb œ œb

p

5

6

œ œb œ œb œ œ œµ œµ œ œb œ œm œn

7

6

œb

œ

p

œ

p

J

œb

R

œ

J

ϵ

R

ϵ3

3

R

œœ

J

œb œœ

œ

π

Œ

7

5

&

œn

f

œ œbœ œ œ œb œ œb

57

. œ œ œ

œ

p

œ# œnœ œb

5

Œ ‰

œ œb œ œµ œ œb

7

œb œ œm œn œ œµ œµ œ œb œ

55

Ó

œm œnœµ

œ ‰

6

Œ ‰

œ œb œ œm œ œb

7

œb

F

°

œ œm œn œ œm œm œ œb œ

55

Ó

œm œnœm

œ

p

6

œµ œµ œ œb œ œm œ œµ œµ œ œb œ œm œn

p

56

œ œn œ œµ œµ œ œb œ œb œ

ϵ

p

3

5

ϵ

f

œ œb œ œb œ

œµœ

j

œ

‰ .

p

R

O

œ

mm

p

5

3

œŒ

ø

J

œ

p

&

œ œ œb œ œm œn œ œBœ œb œ

p

‰ .

3

56

‰ .

œm

p

œ œb œ œµ œm

f

œ œb

4:5

œ

π

.œm œ œ .œb .œ

3 5 7

œ œ

f

œb œ.œ .œm œn œ

œb

œ

p

⋲ .

J

œ

p

œ .œbB œ.œbB

œ

œ

p

œ

p

.œb œ

œ

œµ œ .œ

f

œœ# œ œb œ œb œ œ

f

œbœ œb œ

5:6

5

œ œµ œµ œ œb œ œm œn œ œµ

6

Ó

œµ œ œb

œ œm œnœµ

7

œ

°*

œm œm œ œb œ œm œn œ œm

F6

Ó

œm œ œb

œ œm œnœm

7

p

œb œ œµ œ œb

5 3

‰ .

R

ϵ

p

œ œ œb œ œb

5

B

O

œO

œ

O

œ O

œ

f

.œ.œµ

.œ.œ

F

?

42 Tear of the Clouds

3/16/08

A. Tear of the Clouds score 150

Page 161: Klingbeil Dissertation Web

&

&

&

B

&

&

B

&

?

&

15

&

?

&

&

15

&

?

&

B

?

?

4

4

4

4

4

4

4

4

8

3

8

3

8

3

8

3

Fl.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

182 œb

p

œœ œµ

f

œm œÂ œn œ œµ œÂ

37

182

œ œB œ œµœB œb

p

‰ R

œ

P

œµ œm œ œ

6:5

182œm œ œ œb

f

œb œ œb œ œµ œµ œB œb3

5

182

Œ

‰.œµ

p

œµ œbœm

53

?

182

œ

f

œbœm œ œb

3

182

œ œ

f

œ œb œ œ .œ œ

6

182

œ

œ œ œB œœb

?

182

Œ ‰ .R

œb œœ œb œ

œb œ œ œ# œn œ œb œbœb

p

Œ Œ

67

182

œb œ

Ó

c7

Œ

œœm œ œ œµ œ œB

ϵ

3

5

œœb

œµ œµ œB

Œ Œ

182

œb œ

Ó

Œ

œ

°*

œm œ œ œm œ œœm

F

3

5

œœb

œm œm œn

π

Ó

182œm

f

œ œ œµ œ œB œ œµœB

ϵϵ ϵ

p

œb

p

5

6

œ

f

œ œµ œµ œBœµ œB œB œµ œm œn œ œb

p

5:6

53

œœµ œb

œµ œ œB œµ œÂ

53

œœ

p

œB œµ&

œ œB œ œµ œµ œb

p

‰ .R

œb

f

œœ œµ œm

6

R

œÂ œB

f

œ œ œ œb œbœ

‰ Œ

7:5

œ œ œb œ œb œb œ œ

π

⋲J

œ

f

œœµ

œ8:5

3

œ œµ œµ œm œn

f

œœm

œ œm œm œn œ

p

5

œœm œn œ œb œb

π

⋲ .œ

p

3

œ.œm

.œ œb œb œ œm

r

œn

p

R

œ

P

.ϵ.ϵ ϵ

j

œB

p

5

œ œ œœ

œ œ

f

œ# œb œb

œb œn

⋲Œ

5

7

Œ

œ œ# œn œb

œb

p

7

Œ ‰ .r

œb œœ

œm œ œ

5

œµ œbœµ

ϵϵ

5

œm œ œb œBœb

œµ œB œB

œ

‰ .

7

Œ ‰ .r

œb

p

œœ

œm œ œ

5

œm œbœm

œ

œm

5

œm œ œb œœb

œm œn œ

œ

π

‰ .

7

œ œœ œµ œm

f

œ œ œµ œÂ œnœ œµ œµ7

5

Œ

œ

p

œœ œµ

œµ œm œnœm œÂ œb

œµ œB œB

p

⋲ ⋲

.

J

œ

p

6

œm œn œ

œb

œ œµ œµ œ œµ

f

3

7

5

?

q = 80

q = 80

œ

œm œ œµ œB œBœµ œ

p

⋲‰ œm

f

œ œµ œB œBœµ œ

œm œ

5

3

6

Œ ‰ œm

f

œ œµ œB œBœµ

œm œ œµ œB

7

œm œ œµ œB œBœµ œµ œb œb œ œm œ œb

p

.œm

f3

5

5

ϵ

p

œµ œb œb œB œ

ϵ

f

œµ œb œb œB œ œ œb

35

œbB.œ

ff

.œb

j

œb œœb œb

3

œœm

æ

œ

æ

Ͼ

f

œbæ

Ͼ

3

ϵ

p

œµ œb œBœb

œm œB œb

3

5

œ#

p

œ œœ# œ œ œ# œ œb

‰ œœ œ œ œ œ œ

Œ

6 66

Ó Œ

œœ œb œb

œb œ

6

c8

œm œ œB œBœµ œµ œµ œB œµ

Ó

5

Ó

œm œB œµ œ œbœB œb

œ

œb œ œb

5

6

œm

°*

œ œ œ œœm œm œ œm

5

Ó

œm œn œm œ œbœ œb

œ

œb œ œb

5

6

œµœÂ œB œµ œm

œ>

sul E

p

œ

œm>

sul A

œb

J

œ>

sul G

5

œB

f

œµœµ œB œb

œµ œ œb œb

‰ .R

œ

sul A

3

&

.œµœµ œB

f

œb œB œµ œ œb œb œB œb.

œB.

œ. œµ

.

5

‰J

ϵ

p

œ œb œb œB œb œB œ. œµ

.œB

.œB

.3

3

43Tear of the Clouds

3/16/08

A. Tear of the Clouds score 151

Page 162: Klingbeil Dissertation Web

&

&

&

?

&

&

?

&

?

&

15

&

?

&

&

15

&

?

&

&

?

?

8

3

8

3

8

3

8

3

4

2

4

2

4

2

4

2

4

3

4

3

4

3

4

3

4

4

4

4

4

4

4

4

Fl.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

Ÿ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~185

œµ œB œBœµ œ œ

œm( )J

œ

6

185

œBœµ

Œ

185

J

œ œµ œB œB

3

185

œm œn œb œ œœb

6

B

185

œ œ œ œb

r

œb

p

185

œb

æ

œ

æ

œ

æ

p

3

185

œB

>

f

œ

> ϵ

>

r

œ

5

185

.Ͼ

f

185

‰ œ œ œ œn

œ

œµœB

œ

‰ . ‰

7

185

‰ œm

f

œ œ œ

œ

œmœ

œ

‰ . ‰

7

185

œmJ

œ>

ƒ

œmJ

œ

>

œ

r

œ

>œm

œb ‰

ϵ.

ϵ.

œB.

œb.

f

q = 100

q = 100

p

˙

œm œB œb

r

œ

p

‰ .3

œ>

f

œµ œœµ

œm œB œbœ

?

Œ

œB

ƒ

œ>

f

.ϵ>

c9

œm œµœÂ œ

‰œ œ

œµœm œB ⋲

œµ œœµ

œm œB

5

œm

°*

œnœ

œ

‰œm œn

œ

œm œ ⋲

œm œ œ

œm œ

5

Ó

.œ>

sul E

ƒ

œ

J

œ>sul A

œm>

ƒ

sul D

œ

.

J

œ>

B

œœ

œm

⋲ œµ œµœB

Œ

œ

pizz.

ƒ

Ó ⋲

œ

f

œB œm

œœ

œB

œb>

f

œBœB

œ œ œ œBœm œB

œ

œœb

œ

>

œb>

œœ

œ

œœ

>

œb

>œb

>

œb

>

j

œb

>

‰?

Œ

œ˘

ƒ

œm˘

œn˘

œ

˘

œm

>

œ œ œœ

p

5

.œm

> .œb

>

œ>

œ

˘

ƒ

3

œÂ œ œµ œœµ

Œ ‰œ œB

5

Œ

œm œB œbœ

œ

œm

œœµ œ

Œ

6

7

œœ œm œ œ

Œ ‰œ œ

5

Œ

œm œ œbœ

œ

œm

œœm œ

Œ

6

7

j

œb

.

J

œ>

sul D

œm

.

J

œm>

sul G

œ

J

œm>

œb

J

œm>

R

œ

œ>

sul G

œb

.

j

œ

>

sul C

œ

.

j

œ

> œ

œbœ

œ

œm

ϵ

sul pont.

ƒ

œœ

œ

œ

œm

œœµ œ

3

œ>

Ï

œ>

ϵ>

œ˘

œb˘

œb˘

œm˘

œ˘

œœµ œ œB œB œ œ

f

5

7

‰ .œ

ƒ

œmœb

œ

œœµ œ œb œb

œmœ

œµ œ œ

76

ϵ>

ƒ

œ œb œbœm œµ œ œ

œ œ œB

>

œ œ œm œµ œ œm6

7

œ>

f

œÂ œ œ œm œµ œ œm œm œB

>

œ œbœ œ

œm5

6

˙b

>

Í

Œ

ƒ

˙

>

Í

Œ

Ï

?

˙µ>

Í

Œ

ƒ

∑ ÷

c10

œ

œ œBœB

œ œ œm œµ œ

‰ Œ

5

6

Œœm œµ œB œB œ

œb œB œ œb

œµœ

ϵ

5

7

t

œ

Ï

œ œ

œœ œ œm œn

œ

‰ Œ

5

6

Œœm œn œ œ œ

œb œ œ œbœb

œœm

5

7

t

œb

‰œ

ƒ

œµ œ œBœ

œœµ œ œB

œ œ œm œB œb œb3

6

6

.

J

œ

ƒ

œµ œ œBœ

œmœ

œm œB œB œ5

6

œµord. œ œBœ

œµœm œB

œB œ œb œb œB œ œb

5

6

Œ

œb

æ

arco

œb

æ

œB

æ

œ

æ

œb æ

ϵ

æ

æ

œ

æ

44 Tear of the Clouds

3/16/08

A. Tear of the Clouds score 152

Page 163: Klingbeil Dissertation Web

&

&

&

?

?

?

?

÷

&

15

&

t

&

&

15

&

t

&

B

?

?

4

4

4

4

4

4

4

4

4

3

4

3

4

3

4

3

4

2

4

2

4

2

4

2

4

3

4

3

4

3

4

3

4

4

4

4

4

4

4

4

Fl.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

Ÿ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~

Ÿ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~

Ÿ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~

q = 69

q = 69

F

F

189

.œm

,

Ï

j

œm ˙

jetwhistle

189

189

j

œm

^

,

Ï

‰ Œ Ó

189

˙m

^

ƒ

,

˙

189

œb

^

ƒ

,

189

˙

ƒ

,

˙

(pedal)

189

˙m

ƒ

,

˙

189

œ

>

f

,

Œ Ó

bass drum

189

wm

189

wm

189

j

œm

,

Ï

‰ Œ Ó

.˙n

Ï

,

œ

p

.˙m

ÿ

Ï

,

œ

p

œ

Œ

ø

‰j

œ

ø

œ

œb( )j

œb

ø

˙

Ó

ø

Œ

ø

œ

j

œ ‰

ø

Ó

j

œ

ø

œ œ

f

Ͼ

sul pont.

˙æ

œ œ

f

Ͼ

sul pont.

˙æ

ø

f

Œ ⋲.

j

œ

ø

œ

F

œ( )j

œm

ø

˙

œB( )j

œ

ø

˙

.˙æ

.˙æ

ø

˙

˙

∑&

∑&

˙

˙

˙æ

˙æ

p

˙

Œ

ø

Œ

˙

p

Œ Œ

j

œB

p

œ

3

J

ϵ

p

˙

f

sul pont.

f

sul pont.

Œ

˙b

ø

ord.

.˙m

p

ord.

Ó ˙B

ø

w

F

Ó

˙

ø

w

w

w

.˙m

œ

œ

j

œ

ƒ

˙

j

œ

ƒ

˙

45Tear of the Clouds

3/16/08

A. Tear of the Clouds score 153

Page 164: Klingbeil Dissertation Web

&

&

&

?

&

&

?

÷

&

15

&

t

&

&

15

&

t

&

B

?

?

4

3

4

3

4

3

4

3

4

4

4

4

4

4

4

4

4

3

4

3

4

3

4

3

4

4

4

4

4

4

4

4

Fl.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

ø

ø

195

œ

f

195j

œ

ø

‰ Œ Ó

195œ

F

195

195

˙

ø

.

j

ø

air only

195

˙Ó

ø

take mute

195

˙ ⋲

ø

.

J

ø

air only

195

195

Ó

œœœœœœœœ

œ

µ

m

b

µ

m

ø

195

195

œ

P

ΠO

˙

ø

œ

P

Œ

O

˙

µµ

ø

.˙b

>

œ

p

>

œ

p

p

ƒ

≠ ≠

ƒ

J

p

O

˙

p

O

œ

O

˙

p

O

œ

∑&

q = 66

q = 66

Ó Œ ⋲ .

J

œB>

Íp

Ó Œ œb

>

Íp

Ó ‰ .r

œm

>

Íp

œ

Ó ‰ J

œm>

Íp

œ

Ó

˙µ

Íp

Íp

œ

œœœ

œ

m

b

Â

F

œ

œœœ

œ

m

b

m

F

Œ Ó

O

w

µµ

p

O

w

p

j

œB

Íp

w

o

p

w

w

w

w

w

œm>

con sord.

f

.œn ˙ .œ

®

œm>

P

w

w

œ

.œn ˙

5

∑&

Œ

˙

p

∑?

46 Tear of the Clouds

3/16/08

A. Tear of the Clouds score 154

Page 165: Klingbeil Dissertation Web

&

&

&

?

&

&

?

&

&

15

&

t

&

&

15

&

t

&

B

?

?

4

4

4

4

4

4

4

4

4

3

4

3

4

3

4

3

4

4

4

4

4

4

4

4

Fl.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

200

.˙ Œ

200

.˙ Œ

200

.˙ .œ œÂ

>

Íp

200˙

˙m>

Íp

200

Ó Œ

œB

>Íp

200

œ

π

Œ Ó

200

Ó ‰ J

œB>

Íp

œ

200œ#

p

Œ Ó

crotales

200

œ

œ

œB

µB

œ

ϵ

F

200

œ

œ

œ

m Œ Ó

œ

œm

F

Œ Ó

200

œ

Œ

ø

Œ ‰

J

œb>

Íp

.

J

ϵ>

Íp

œÂ>

Íp

3

w

>

Íp

w

w

w

Ó Œ

.œm

p

œn

w

Œœ

Ó

w

w

Œ ⋲

.

J

œ

Œ

&

∑&

Œ

.˙µæ

ø

flz.

˙B>

Ï

˙

œ

f

œœµ

œ œm œ

œm

fl

Ï

Œ Ó

ϵ

f

œm œnœ œm œb

œ

fl

Ï

Œ Œœm

p

.˙µ

ƒ

j

œ

p

>

Sf

j

œ

p

œœœµ

µB

œ

œœœ

B

œ

œ

œ

œœœb

Â

m

m

ƒ

œœœb

ƒ

Œ Ó

O

w

µµ

æ

ƒƒ

O

w

æ

ƒƒ

O

w

bb

æ

ƒƒ

œœ

ƒ

sul pont.

œœ œœ

ord.

F

˙˙

˙æ

˙æ

ƒ

w

pp

w

Ó Œ ⋲

.

j

œb

p

∑ ÷

O

w

æ

O

w

æ

O

w

æ

ww

47Tear of the Clouds

3/16/08

A. Tear of the Clouds score 155

Page 166: Klingbeil Dissertation Web

&

&

&

?

&

&

?

÷

&

15

&

t

&

&

15

&

t

&

&

&

?

4

2

4

2

4

2

4

2

4

4

4

4

4

4

4

4

Fl.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~

205

.˙æ

ϵ

ø

205œ œ .˙

205

205

w

205

205

˙

j

œ œ .œ œ

3 5

205

205

Œ

π

sandblocks

205

205

205 O

w

µµ

p

O

w

p

O

w

p

ww

ff

sul pont.

p

w

w

Œ

.˙m

π

œ

π

J

œ‰ Ó

œm œœ œ

œB>

F

3

Ó

P

O

w

Ó

˙B

ø

O

w

w

p

w

F

w

P

Ó

sus cym. – brass mallet

w

F

sul pont.

w

ord.

œ

Œ

ø

Ó

w

π

j

œ

ø

‰ ‰ j

œ

π

˙

p

ƒ

Œ

air only

Ó Œ

œB>

p

p

ƒ

Œ

air only

r

Ó

>

f

‰ .

J

œ

π

‰ Œ Ó

j

œ

ø

‰ Œ Ó

ø

ø

w

w

œ œ

.œB>>

œ œ

5

Ó ‰

j

œ

F

pizz.

Œ

˙

∑&

∑ B

∑?

Œ

J

œ⋲

.

J

œÂ

f

3

48 Tear of the Clouds

3/16/08

A. Tear of the Clouds score 156

Page 167: Klingbeil Dissertation Web

&

&

&

?

&

&

?

&

&

15

&

t

&

&

15

&

t

&

B

?

?

4

4

4

4

4

4

4

4

Fl.

Ob.

Cl.

Bsn.

Hn.

Tpt.

Tbn.

Comp.

Vn.

Vla.

Vc.

Cb.

Kbd.

Synth

Perc

q = 60

q = 60

211

R

œ˘

S

‰ . Œ Ó

211

211

211

r

œ

fl

S

‰ . Œ Ó

211

211j

œ‰ Œ Ó

211

211œ

f

Œ Ó

crotales

÷

211œ

Œ Ó

œm

°

Ï

211

œm

ƒ

Œ Ó

211

Ó Œ

h

h

@

p

bow directly on bridge(noise only)

Ó Œ

hhm@

p

bow directly on bridge(noise only)

Ó Ó

wm

>

ƒ

arco

Œ ‰

j

ø

≠ ≠

f

air only

Ó

ø

air only

Œ

ø

≠ ≠

f

air only

Ó Œ

ø

air only

Ó Œ œ

p

medium yarn sus. cym.dome

h

h

@

hh

@

Œ

.

.hh

@

p

bow directly on bridge(noise only)

w

ø

Œ Ó

f

≠ ≠

ø

fl

f

fl

fl

fl

≠ ≠

ø

fl

f

fl

fl

fl

fl

fl

fl

fl

≠ ≠

f

≠˘≠˘≠˘≠˘≠˘≠˘≠˘≠

˘≠˘≠˘≠˘

h

h

@h

h

@ Ó

ø

hh

@hh

@ Ó

ø

.

.hh

@ Œ

ø

œ

ø

Œ Ó

fl

fl

fl

fl

fl

fl

fl

fl

ø

Ó

fl

fl

fl

fl

fl

fl

fl

fl

ø

Ó

≠˘≠˘≠˘≠˘≠

˘≠˘≠˘≠˘

ø

Ó

12"

12"

12"

12"

12"

12"

12"

12"

12"

12"

12"

12"

12"

12"

12"

fade out

March 14, 2008New Haven, Connecticut

49Tear of the Clouds

3/16/08

A. Tear of the Clouds score 157