8/11/2019 Dissertation New
1/97
ADAPTIVE FOURIER ANALYSIS FOR UNEQUALLY-
SPACED TIME SERIES DATA
By
Hong Liang
Dissertation submitted to the faculty of the
Virginia Polytechnic Institute and State University
in partial fulfillment of the requirements for the degree of
Doctor of Philosophy
in
Statistics
Robert V. Foutz, Chair
Marion R. Reynolds, Jr.
Donald R. Jensen
George R. Terrell
Christine Anderson-Cook
April 16, 2002Blacksburg, Virginia
Keywords: Adaptive Fourier Analysis, Unequally-Spaced, Time Series, Walsh-Fourier,Wavelet
8/11/2019 Dissertation New
2/97
Adaptive Fourier Analysis For Unequally-Spaced Time SeriesData
by
Hong Liang
Robert V. Foutz, Chairman
Statistics
(ABSTRACT)
Fourier analysis, Walsh-Fourier analysis, and wavelet analysis have often been used in
time series analysis. Fourier analysis can be used to detect periodic components that have
sinusoidal shape; however, it might be misleading when the periodic components are not
sinusoidal. Walsh-Fourier analysis is suitable for revealing the rectangular trends of time
series. The flaw of the Walsh-Fourier analysis is that Walsh functions are not periodic.
The resulting Walsh-Fourier analysis is more difficult to interpret than classical Fourier
analysis. Wavelet analysis is very useful in analyzing and describing time series with
gradual frequency changes. Wavelet analysis also has a shortcoming by giving no exact
meaning to the concept of frequency because wavelets are not periodic functions. In addi-
tion, all three analysis methods above require equally-spaced time series observations.
In this dissertation, by using a sequence of periodic step functions, a new analysis
method, adaptive Fourier analysis, and its theory are developed. These can be applied to
time series data where patterns may take general periodic shapes that include sinusoids as
special cases. Most importantly, the resulting adaptive Fourier analysis does not require
equally-spaced time series observations.
8/11/2019 Dissertation New
3/97
iii
Acknowledgements
I would like to sincerely express my appreciation to Dr. Robert V. Foutz, my aca-
demic advisor, who introduced me to time series analysis. His excellent guidance, cour-
ses, support, wisdom, patience, and enthusiasm were essential throughout my academic
career at Virginia Tech.
I would like to thank all of my committee members: Dr. Reynolds, Dr. Jenson,
Dr. Terrell, and Dr. Anderson-Cook. I want to thank Dr. Reynolds for his wonderful
courses helping me intuitively understand applied statistics and academic research me-
thods. I would like to send thanks to Dr. Jensen for his careful reading of my dissertation.
To Dr. Terrell, I thank him for his comments and suggestions. Lastly, I would also like to
thank Dr. Anderson-Cook for her many suggestions of modifications in my dissertation.
I want to give my thanks to all faculties at Virginia Tech in the Statistics Depart-
ment, Graduate School, and International Student Center. In the Statistics Department, I
would especially like to thank Dr. Good who offered me a reference paper for my disser-
tation. In Graduate School, I would especially like to thank Dr. McKeon, for his outstand-
ing service for international students.
Finally, I would like to thank my wife, Haiming and my son, Sixing for their love
and support.
8/11/2019 Dissertation New
4/97
iv
Table of Contents
1 Introduction 1
2 Review of data analysis methods 4
2.1 Fourier Analysis . 4
2.2 Walsh-Fourier Analysis 7
2.3 Wavelet Analysis . 12
2.4 The spectral analysis methods for unequally-spaced time series data 17
2.4.1 Interpolation methods ..... 17
2.4.2 Least-Squares methods 19
2.4.3 Linear algebra method and the others .. 21
3 Adaptive Fourier Analysis 23
3.1 Motivation 23
3.2 The periodic step functions 24
3.3 The properties of the periodic step functions 27
3.4 Multiresolution analysis 28
3.5 The adaptive Fourier analysis of the digital time series .. 29
3.6 Methodology .. 32
4
Examples 36
Example 4.1 .. 36
Example 4.2 . 45
Example 4.3 . 49
8/11/2019 Dissertation New
5/97
v
Example 4.4 . 52
Example 4.5 54
5 Conclusion and future research 58
5.1 Summary and conclusion . 58
5.2 Proposed future research .. 59
Bibliography 61
Appendix 64
Appendix 1. Proof of Theory 3.1 and its Corollary 3.1 .. 64
Appendix 2. The construction of x e (t) . 73
Appendix 3. Proof of x(t) =n
limm
lim s mn, (t) 75
Appendix 4. A nonstandard inner product < y, z> 4 77
Appendix 5. Proof of (3.12) . 77
Appendix 6. Background .. 79
A6.1 Vector space . 79
A6.2 Inner product space, Hilbert space and their properties 82
A6.3 Orthogonal projection and orthonormal bases 87
Vita 90
8/11/2019 Dissertation New
6/97
vi
List of Figures
Figure 2.1 The first eight Walsh functions . 10
Figure 2.2 Some basis functions for Fourier transform .. 11
Figure 2.3 Four different mother wavelets ..... 13
Figure 2.4 The signal ss = 10*cos(pi*t/15)+3*cos(pi*t/10) ... 15
Figure 2.5 Decomposition of SS into the sum of 16 Wavelet functions . 15
Figure 2.6 Doppler(t) =((t*(t-1))**0.5)*sin(2.1*pi/(t+0.5)) 16
Figure 3.1 The periodic step function in the complex plan 25
Figure 3.2 The step function f )(,6,2 tj ..... 26
Figure 4.1 Monthly accidental deaths in the U.S.A., 1973-1978 .. 36
Figure 4.2 Frequency component, k=6 in Ex. 4.1 ... 38
Figure 4.3 Adaptive Fourier line spectrum for Ex. 4.1 ... 39
Figure 4.4 Fourier line spectrum in Ex. 4.1 39
Figure 4.5 Monthly accidental deaths without Mar. in Ex. 4.1 40
Figure 4.6 Frequency component, k=6 in Ex.4.1 without Mar. ... 40
Figure 4.7 Monthly accidental deaths without Jan., , Jun., respectively .. 42
Figure 4.8 Frequency component, k=6, without Jan., , Jun., respectively ... 42
Figure 4.9 Monthly accidental deaths by deleting randomly 10 points, first time 43
Figure 4.10 Frequency component, k=6, deleting randomly 10 points, first time... 43
Figure 4.11 Monthly accidental deaths by deleting randomly to points, second time..44
8/11/2019 Dissertation New
7/97
vii
Figure 4.12 Frequency component, k=6, deleting randomly 10 points, second time.. 44
Figure 4.2(1) Original time series in Ex.4.2 . 46
Figure 4.2(2) k = 2 freq. Component in Adaptive Fourier analysis 46
Figure 4.2(3) k = 2 freq. component(-) in Fourier analysis & original signal(.) 47
Figure 4.2(4) Multiresolution decomposition of the signal 48
Figure 4.2(5) Time-scale plot for the signal .. 48
Figure 4.3(1) Original time series in Ex.4.3 .50
Figure 4.3(2) Reconstruction of the time series in Ex.4.3 .50
Figure 4.3(3) Adaptive Fourier ANOVA in Ex.4.3 ...51
Figure 4.3(4) Fourier ANOVA in Ex.4.3 .. 51
Figure 4.3(5) Frequency component k=3 in Ex.4.3 51
Figure 4.3(6) Frequency component k=2 in Ex.4.3 51
Figure 4.4(1) Original time series in Ex.4.4 ...53
Figure 4.4(2) Reconstruction of the time series in Ex.4.4 .. 53
Figure 4.4(3) Adaptive Fourier line spectrum in Ex.4.4 53
Figure 4.4(4) Frequency component k=3 in Ex.4.4 54
Figure 4.4(5) Frequency component k=2 in Ex.4.4 54
Figure 4.5(1) Original time series in Ex.4.5 56
Figure 4.5(2) Reconstruction of the time series in Ex.4.5 ... 56
Figure 4.5(3) Adaptive Fourier line spectrum in Ex.4.5 . 57
Figure 4.5(4) Fourier line spectrum in Ex.4.5. 57
Figure 4.5(5) Frequency component k=2 in Ex.4.5.57
Figure 4.5(6) Frequency component k=1 in Ex.4.5 . 57
8/11/2019 Dissertation New
8/97
1
1 Introduction
In the statistical analysis of time series, Fourier analysis has provided a general method for
discovering or analyzing the periodicity and examining the global energy-frequency dis-
tribution in time series data. It has been a valuable and powerful tool of time series analy-
sis, but it still has some crucial limitations : the data must consist of equally-spaced obser-
vations; and periodic components in the data must be sinusoidal . Otherwise, the analysis
might give misleading results.
For non-sinusoidal waveforms, such as square-wave and rectangular waveform, Fourier
analysis needs too many additional harmonic components to decompose approximately
the waveform functions, and therefore it spreads the energy over a wide frequency range,
which might cause Fourier ANOVA to be misunderstood. Beauchamp (1975,1984) empiri-
cally demonstrated that when a time series is based upon a sinusoidal waveform, Fourier
analysis is more efficient, and when a time series is rectangular or with sharp discontinui-
ties, Walsh-Fourier analysis is more efficient. However, the Walsh functions are not perio-
dic, and therefore the Walsh-Fourier decomposition is more difficult to interpret than the
Fourier harmonic decomposition.
In actual time series analysis, given a sequence of N values x =(x1, x2, , xN), by using the
discrete Fourier Transform (DFT), x can be evaluated only at a special set of N/2 evenly
spaced frequencies . When x comes from a non-sinusoidal function containing only less
than N/2 frequencies, DFT may cause energy spreading.
8/11/2019 Dissertation New
9/97
2
For the time series of unequally-spaced data, Fourier analysis, the Discrete Fourier Trans-
form, cant be used directly. A variety of methods have been suggested to overcome this
limitation of the DFT due to unequally-spaced data. Among them are the interpolation me-
thod, the Least-squares method, and the algebra method. But these methods have some
shortcomings, and they also suffer the shortcoming of Fourier spectral analysis , since
they are all Fourier based.
In wavelet analysis, a given function x(t) is decomposed into a sum of wavelet functions.
These wavelet functions are derived from a single mother wavelet function (t) by apply-
ing varying translations and dilations to (t). Given a mother wavelet function (t), for
real a>0, and real b, a sequence of wavelet functions can be derived as
ba, (t) = a21 (
abt ),
where arepresents the scale parameter and brepresents the translation parameter. In a
loose sense, we may identify the wavelet parametersa, b
with frequency and time ,
respectively. See Priestley(1996). Wavelet analysis is very useful in analyzing and describ-
ing time series data with gradual frequency changes. However, wavelets are not periodic
functions, and the concepts of frequency and of periodicity have no precise meaning in the
resulting wavelet analysis. In addition, different mother wavelet may have different analy-
sis results.
In this research, we develop a new analysis method, adaptive Fourier analysis and its
theory. The advantage of adaptive Fourier analysis and its theory is that it can be applied
8/11/2019 Dissertation New
10/97
3
to time series data where patterns may take general periodic shapes that include sinusoids
as special cases. Most importantly, the resulting adaptive Fourier analysis does not require
equally-spaced time series observations.
Chapter 2 reviews the present available time series data analysis methods. In Chapter 3,
adaptive Fourier analysis and the definition of ADFT are presented. Chapter 4 gives some
examples. Chapter 5 makes conclusions and suggestions for future research. Proofs are
contained in the Appendix1-5. Appendix6 provides the background for theoretical results
and their proofs in this research. All conceptions and their properties used in Chapter3 and
Appendix1-5, such as vector space, direct sum, inner product, Hilbert space, weight inner
product, orthogonal projection, and Gram-Schmidt procedure, are presented in Appendix6.
8/11/2019 Dissertation New
11/97
4
2. Review of data analysis methods
2.1 Fourier Analysis
2.1.1 Fourier Series
The basic idea of a Fourier series is that any function x(t) L2[0,T] can be decomposed
into an infinite sum of cosine and sine functions:
x(t)= ]2
sin2
cos[0 T
ktb
T
kta
k
kk
=
+ , for all t (2.1)
where ak= T
txT 0
)(1
cosT
kt2dt , (2.2)
bk= T
txT
0
)(1
sinT
kt2dt. (2.3)
This is due to the fact that {1, cosT
kt2, sin
T
kt2, k=1, 2, 3, ...} form a basis for the space
L
2
[0,T]. The summation in (2.1) is up to infinity, but x(t) can be well approximated in
the L2sense by a finite sum with K cosine and sine functions:
X(t) ]2
sin2
cos[0 T
ktb
T
kta
K
k
kk
=
+ . (2.4)
This decomposition shows that x(t) can be approximated by a sum of sinusoidal shapes at
frequencies k = 2k/T, k = 0,1, , K. In addition, the variability inx(t) as measured by
dttx
T
0
2|)(| can be approximately partitioned into the sum of the variability of the sinusoi-
dal shapes:
8/11/2019 Dissertation New
12/97
5
dttxT
0
2|)(| = T
0
[ ]2
sin2
cos[0 T
ktb
T
kta
K
k
kk
=
+ 2dt 2
0
||=
K
k
k . (2.5)
A standard technique of time series analysis is to treat the partition (2.5) as an analysis of
variance (ANOVA) for identifying sinusoidal periodicities in a time series data set {x(t),
0< t T}. When x(t) has sharp discontinuities or a non-sinusoidal waveform, such as a
rectangular waveform, then we would require a very large number, K, of terms in its
Fourier series in order to get an adequate approximation.
2.1.2 Discrete Fourier Transform (DFT)
For an arbitrary time series data set,x = (x(t 1 ),x(t 2 ), ,x(tN ) ), if the observation times
are equally spaced, at interval t, then the data set x can be simply written asx= (x(1),
x(2), ,x(N) ) by taking t =1 and t j =j; and there is an orthogonal system { ekitw : - 2
N +
1 k 2N
if n is even and - 21N
k 21N
if n is odd }, so that the Discrete Fourier
Transform ofxcan be defined by
X*(k) =
N
1
=
N
t
itwketx1
)( , (2.6)
where the frequencies wk=2k/N, k=0,1, , 2/N , are called the Fourier frequencies;
and r is the largest integer no larger than r. The Fourier series ofx(t) can be written as
x(t) =
=
2/
2/)1(
* )(N
Nk
kX e kitw . (2.7)
Corresponding to (2.5), we have the ANOVA
8/11/2019 Dissertation New
13/97
6
NtxN
t
==
)(1
2
=
2/
2/)1(
2* |)(|N
Nk
kX . (2.8)
This representation provides an ANOVA for revealing how well the periodicities in x
may be described by the sinusoidal shapes X*(-k)e kitw + X*(k)e kitw . The ANOVA
decomposition in (2.8) holds only if the DFT, X* (t), is evaluated only at a fixed set of
N/2 equally-spaced frequencies wk, and the data set must be equally-spaced.
2.1.3. Spectrum ANOVA for Equally-Spaced Time Series Data
Let L 2be the set of all continuous-time, complex-valued functions. L 2 ={y(t), 0 t N}
for which the Lebesque integral N
ty0
2|)(| dt is finite. The set L 2 is a Hilbert space with
inner product
1 =N
1
N
tzty0
)()( dt, (2.9)
where )(tz is the complex conjugate of z(t). Let A k be the subspace of L2 that contains
all periodic functions that have the form c*exp(i k t) for complex-valued scalars c and for
k = Nk2 . Also, let r be the largest integer no larger than r. Classical analysis shows
that corresponding to every equally-spaced digital times series data set x = (x(1), x(2),,
x(N) ), there is a continuous-time function s in the N-dimensional subspace,
L NA, = A ][2
1 N+ + A
][2
N , (2.10)
8/11/2019 Dissertation New
14/97
7
with the property that s(t) = x(t) at t = 1, 2, , N. See Koopmans ( 1974, P. 17). The
function is s = P 2 ( x | L NA, ), the projection of x onto L NA, with respect to the inner
product
2 =N
1
=
N
t
tzty1
)(*)( . (2.11)
Because the subspaces A k in (2.10) are orthogonal with respect to the inner product < y,
z> 2 , the function s = P 2 ( x | L NA, ), has the Fourier series representation,
s(t) =
)|(
2/
2/)1(2
=
N
Nk
kAxP (t), (2.12)
where P 2 (x | A k ) is the projection of x onto A k with respect to < y, z > 2 . The corres-
ponding partition of |s| 22 = 2 is
|s| 22 =
2
2/
2/)1(2 |)|(|
=
N
Nk
kAxP . (2.13)
In the Fourier series (2.12), each frequency component P 2 ( x | A k ) + P 2 ( x | A k ) is a
sinusoid with frequency k . The corresponding partition (2.13) provides an ANOVA for
revealing how well the periodicities in s ( and in x ) may be described by the sinusoidal
shapes P 2 ( x | A k ) + P 2 ( x | A k ).
2.2 Walsh-Fourier Analysis
Fourier series can be used to partition functions into sinusoidal waves. Unlike Fourier
analysis, Walsh-Fourier analysis deals with the decomposition of functions into the
8/11/2019 Dissertation New
15/97
8
Walsh functions, which are rectangular waves. These waves have been used in signal
transmission, speech analysis and pattern recognition, etc. Walsh-Fourier analysis is
suited to the analysis of discrete valued and categorical valued time series. See Stoffer
(1991).
The Walsh functions are defined as products of the Rademacher functions and the Rade-
macher functions are defined as below.
Definition ( Rademacher functions) Consider the function defined on the half open
interval [0,1) by
0 (t) =
.)1,2/1[,1
)2/1,0[,1
tfor
tfor (2.14)
Extend it to the real line by periodicity of period 1 and set k(t) = 0 (2kt) for k = 0,1,
and real t. The functions k(t) are called the Rademacher functions. The Walsh system
{ W[0, t], W[1, t], W[2, t], }is obtained by taking all possible products of Rademacher
functions , where W[0, t] is defined as W[0, t] =1 and the other W[n, t] , n 1 is defined
by
W[n, t] = =
k
i
iit
0
))(( , (2.15)
where n= =
k
i
i
i
0
2 , k=1 and i=0 or 1 for i= 0, 1, ,k-1.
There are a number of definitions of the Walsh functions. Another definition of the
Walsh functions { W[0, t], W[1, t], W[2, t], }is by induction as follows:
8/11/2019 Dissertation New
16/97
9
Initialize the induction by defining
W[0, t] = 1, t [0, 1),
W[1, t] =
).1,2/1[,1
)2/1,0[,1
t
t (2.16)
Then proceed recursively for n =1, 2, , through
W[2n, t] =
8/11/2019 Dissertation New
17/97
10
Figure 2.1 The first eight Walsh functions
0 0 . 1 0 . 2 0 . 3 0 . 4 0 . 5 0 . 6 0 . 7 0 . 8 0 . 9 1
-1
0
1
W[0,t
]
0 0 . 1 0 . 2 0 . 3 0 . 4 0 . 5 0 . 6 0 . 7 0 . 8 0 . 9 1
-1
0
1
W[2,t
]
0 0 . 1 0 . 2 0 . 3 0 . 4 0 . 5 0 . 6 0 . 7 0 . 8 0 . 9 1
-1
0
1
W[4,t
]
0 0 . 1 0 . 2 0 . 3 0 . 4 0 . 5 0 . 6 0 . 7 0 . 8 0 . 9 1
-1
0
1
W[6,t
]
0 0 .1 0 .2 0 .3 0 .4 0 .5 0 .6 0 .7 0 .8 0 .9 1
-1
0
1
W[1,t
]
0 0 .1 0 .2 0 .3 0 .4 0 .5 0 .6 0 .7 0 .8 0 .9 1
-1
0
1
W[3,t]
0 0 .1 0 .2 0 .3 0 .4 0 .5 0 .6 0 .7 0 .8 0 .9 1
-1
0
1
W[5,t
]
0 0 .1 0 .2 0 .3 0 .4 0 .5 0 .6 0 .7 0 .8 0 .9 1
-1
0
1
W[7,t
]
8/11/2019 Dissertation New
18/97
11
-4 -3 -2 -1 0 1 2 3 4-1
0
1
sin(t)
Figure 2.2 Som e basis functions for Fourier transform
-4 -3 -2 -1 0 1 2 3 4-1
0
1
sin(2
t)
-4 -3 -2 -1 0 1 2 3 4-1
0
1
sin(3t)
-4 -3 -2 -1 0 1 2 3 4-1
0
1
cos(t)
-4 -3 -2 -1 0 1 2 3 4-1
0
1
c
os(2t)
-4 -3 -2 -1 0 1 2 3 4-1
0
1
cos(3t)
The Walsh functions, W[n, t] , n1, form a complete orthogonal sequence on the interval
[ 0, 1 ) and they have square wave shapes in which each Walsh function W[ n, t ] takes on
only the value 1 and +1. For any functionx(t) which has period 1 and is Lebesgue inte-
grable on [0,1), it can be decomposed into an infinite sum of the Walsh functions.
x(t)
=0i
aiW[i, t], (2.19)
where ai = 1
0
x(t) W[i, t] dt , i=0,1,2, (2.20)
For equally spaced time series data x= (x(0),x(1), ,x(N-1)) and N is a power of 2, the
8/11/2019 Dissertation New
19/97
8/11/2019 Dissertation New
20/97
13
Another commonly used wavelet is Morlet wavelet defined as
(t) =e2t cos( t 2ln/2 ) e
2t cos(2.885 t). (2.24)
Four different mother wavelets: Haar, Daublet, Symmlet, and Coiflet are shown in Figure
2.3, where the first letter of the wavelet indicates the name: d for Daublet, s for Symmlet,
and c for Coiflet; the number of the wavelet indicates its width and smoothness. See
Bruce and Gao (1996, p.8).
Figure 2.3. Four different mother wavelets
`haar' mother, psi(0,0)
0.0 0.2 0.4 0.6 0.8 1.0
-1.0
-0.5
0.0
0.5
1.0
`d4' mother, psi(0,0)
-1.0 -0.5 0.0 0.5 1.0 1.5 2.0
-1.0
0.0
1.0
`s12' mother, psi(0,0)
-4 -2 0 2 4 6
-1.0
-0.5
0.0
0.5
1.0
1.5
`c12' mother, psi(0,0)
-4 -2 0 2 4 6
-0.5
0.0
0.5
1.0
1.5
Given a mother wavelet (t), an infinite sequence of wavelets can be constructed by
varying translations band dilations aas below
a,b(t) = |a|-1/2
( abt
). (2.25)
By defining the continuous wavelet transform W(a,b) as
W(a,b) = < x(t), a,b(t) > =
x(t)a,b(t)dt , (2.26)
8/11/2019 Dissertation New
21/97
14
we can represent x(t) as
x(t) =1
1
C
0
a-2W(a,b)a,b(t)dadb (2.27)
where C1 = dt
2|)(| and ()=
(t)e-itdt .
When aand btake on discrete sets of values, we can similarly obtain the discrete wavelet
transform as
W(m,n) = =
x(t)m,n(t)dt, (2.28)
and x(t) =
=m
=n
Wm,nm,n(t). (2.29)
For an equally spaced time series data x =(x(1), x(2), , x(N) ), we can take approximate
wavelet transforms by replacing the (2.28) by an estimate such as:
W(m,n) =
x(t)m,n(t)dt
=
N
l
lx1
)( m,n( l ). (2.30)
It follows that a class of discrete wavelet transform (DWT) for equally spaced time series
data can be implemented by using an efficient computational algorithm. See Bruce and
Gao (1996, p37-39).
An example of wavelet approximation is given in Figure 2.5. In the example, the signal
is ss =10*cos(*t/15)+3*cos(*t/10), which is plotted in Figure 2.4. Its wavelet appro-
ximation is plotted in Figure 2.5.
8/11/2019 Dissertation New
22/97
15
Figure 2.4 The signal ss=10*cos(*t/15)+3*cos(*t/10)
Index
ss
0 10 20 30 40 50 60
-10
-5
0
5
10
Figure 2.5 Decomposition of ss into the sum of 16 wavelet functions
D3.3
D3.5
D1.1
D3.4
D3.1
D2.1
D3.8
S5.2
D3.6
D3.2
D5.2
D4.3
D4.2
D5.1
D4.1
D4.4
Approx
0 10 20 30 40 50 60
Another example is a wavelet decomposition for the Doppler signal as shown in Figure
2.6. See Bruce and Gao (p.28). In Figure 2.6, the high frequency oscillations at the begin-
ning of the signal are captured mainly by the fine scale detail components D1 and D2,
while the lower frequency oscillations are captured mainly by the coarse scale compon-
ents D6 and S6.
8/11/2019 Dissertation New
23/97
16
Wavelet analysis is very powerful and efficient in the analysis of data or functions, x(t)
with gradual frequency changes. However, wavelets are not periodic functions. For
example, the Morlet wavelet is Fourier based but its oscillations are dampened by the
exponential factor e2t . In addition, the concepts of frequency and periodicity have no
precise meaning in wavelet analysis. See Priestley (1996).
Figure 2.6 Doppler(t)= ((t*(t-1))**0.5)*sin(2.1* /(t+0.05))
S6
D6
D5
D4
D3
D2
D1
Data
0 200 400 600 800 1000
8/11/2019 Dissertation New
24/97
8/11/2019 Dissertation New
25/97
18
(2.31) can be expressed as a matrix form by
X= Dx , (2.32)
where
X=
)(
)(
)(
1
1
0
NzX
zX
zX
, x =
]1[
]1[
]0[
Nx
x
x
, (2.33)
and
D =
)1(1
21
11
)1(1
21
11
)1(0
20
10
...1
...1
...1
N
NNN
N
N
zzz
zzz
zzz
. (2.34)
If the N sampling points, z 0 , z 1 , , z 1N are distinct, then D is nonsingular, and thus the
inverse of NDFT can be determined by
x = D 1 X. (2.35)
If the N sampling points, z 0 , z 1 , , z 1N are equally-spaced angles on the unit circle in
the z-plane, then ( 2.35 ) corresponds to the classical DFT. X(z) can be expressed as the
Lagrange polynomial of order N-1,
X(z) = ],[
)(
)(1
0k
N
k kk
k zX
zL
zL
=
(2.36)
whereL 0 (z),L 1 (z), ,L 1N (z) are the fundamental polynomials, defined by
L k (z) =
ki
izz ),1(1 k = 0, 1, , N-1. (2.37)
X(z) can also be expressed as the Newton interpolation,
8/11/2019 Dissertation New
26/97
19
X(z) =c 0 +c1 (1-z 0 z1 )+c 2 (1-z 0 z
1 )(1-z 1 z1 )+ +c 1N
=
2
0
1 )1(N
k
kzz , (2.38)
where c 0 =X[z 0 ],
c1 = 110
01
z-1
][
z
czX,
c 2 =)1)(1(
)1(][1
211
20
120102
zzzz
zzcczX,
(2.39)
Note that in (2.39), each jc depends only onX[z 0 ] ,X[z1 ], ,X[z j ] and z 0 , z1 , , z j .
While interpolation methods may be satisfactory in some applications, they all produce
some distortion and loss of information, see Scargle ( 1989 ), and they may cause some
distortion in the spectrum, especially for the data with high frequency components. In
addition, these interpolation methods cant yield an orthogonal and additive spectrum
ANOVA decomposition for the original time series data.
2.4.2 Least-Squares Methods
The least squares method is used on Fourier expansion or the inverse transform to find
the period which minimizes the unexplained variance of the series. This method is very
similar to multivariate regression analysis when multiple periods are present. In the least-
squares sense, a periodogram analysis is based on the trigonometric regression model
x f (t) ==
K
k
PkA
1
( cos(2 t/P k ) + B kP sin(2 t/P k )) + , (2.40)
8/11/2019 Dissertation New
27/97
20
where K is the smallest integer less than or equal to (N-1)/2 and P k = N/k, k = 1, 2, ..., K.
To minimize the mean square difference between (2.40) and the data, one seeks to mini-
mize =
N
j 1 [x(t j )-x f (t j )]
2
, where t 1 , ..., tN are not necessarily equally-spaced, and t N =
N. Thus AkPand B
kPmay be determined by standard linear least-squares techniques.
Some other least-squares methods can be found. The classical DFT power spectrum,
periodogram is defined by (See Brockwell and Davis 1993, p332)
I(k) =N
1| kit
N
t
teX
=
1
|2
=N
1[ 2
1
)sin( tX t
N
t
t =
+ 2
1
)cos( tX t
N
t
t =
]. (2.41)
For unequally-spaced data, t1, t2, , tN are not equally-spaced points. Scargle (1982, 1989),
and Lomb(1976) defined a modified periodogram by
I*() =2
1{
=
=
N
j
j
N
j
jj
t
tX
1
2
1
2
)(cos
)](cos[
+
=
=
N
j
j
N
j
jj
t
tX
1
2
1
2
)(sin
)](sin[
}, (2.42)
where is defined by
tan(2) = j
N
j
t=1
2sin / j
N
j
t=1
2cos . (2.43)
They showed that their modified periodogram and least squares fitting of sinusoidal
waves to the data are exactly equivalent. This method has been called as Lomb-Scargle
method , which has recently been used in biomedical sciences. See Schimmel ( 2001 ),
Van Olofsen, VanHartevelt, and Kruyt (1999), Ruf(1999), Schluz and Stattegger (1997).
8/11/2019 Dissertation New
28/97
21
The least-squares method can detect some frequency components in relatively simple
sinusoidal spectral situations, but for complicated spectra, it encounters more difficulties.
See Swan ( 1982 ). When time series data contain fractions of non-Gaussian noise or con-
sist of periodic signals with non-sinusoidal patterns, Lomb-Scargle method makes more
difficult the interpretation of analysis results, and can lead to misleading estimates of fre-
quency components. See Schimmel ( 2001 ).
2.4.3 Linear Algebra Method
From the classical DFT definition , X
*
(k), the Fourier transform of X(t) is defined by
X*(k) = =
N
t
NktietX
1
/2)( , (2.44)
where k = -(N-1)/2, , (N-1)/2, if N is odd ,
k = -N/2, , N/2, if N is even.
(2.44) can be written as in matrix form by
*X = WNX (2.45)
and X =W1
N
*X . (2.46)
For unequally spaced data uX , Kuhn (1982 ) and Swan (1982 ) defined the DFT expan-
sion similar to (2.46) by
uX = W u1
N
*X , (2.47)
where W u1
N is a matrix function of the unequally spaced data uX and*X is the DFT of
certain unknown equally spaced data X .Substituting (2.45) into (2.46) gives
uX = W u1
NWNX . (2.48)
8/11/2019 Dissertation New
29/97
22
Solving (2.48) gives X , and doing DFT on X gives*X , then , Kuhn(1982) and Paul R.
Swan ( 1982 ) used *X to do spectrum analysis of the data. However, the method is
limited in that the deterministic component of noisy signals must be band limited to less
than the usual Nyquist limit. See Swan (1982).
In addition, the other methods include the string length methods, phase dispersion
minimization method, and the CLEAN algorithm; see Rao, Priestley, and Lessi (1997,
p275-286 ). Regarding the CLEAN algorithm, also see Baisch and Bokelmann. ( 1999).
Other discussions and methods for unequally-spaced time series are in Barthes ( 1995 ),
Engle and Russell (1998), and Good and Doog (1958).
The methods above in Section 2.4.1 through 2.4.3 are Fourier based; they also suffer the
shortcoming of Fourier spectral analysis, i.e., they are not efficient for non-sinusoidal
waveforms, and they might lead to misunderstandings of the spectral ANOVA .
8/11/2019 Dissertation New
30/97
23
3. Adaptive Discrete Fourier Analysis
3.1 Motivation
The purpose of the research is to generalize the Fourier analysis of the digital data x. The
generalization begins by replacing each space A k in the representation of L NA, in ( 2.10)
with a larger space B k ; and replacing an equally-spaced digital time series data set x =
(x(1),, x(N)) in (2.10) with an arbitrary digital time series data set x =(x(t 1 ), , x(tN)),
where t 1 , , tN are not necessarily equally-spaced and tN= N. While each function s in
A k has real and imaginary parts that are sinusoids, the real and imaginary parts of func-
tions in B k have more general periodic shapes. Like sinusoids, these shapes satisfy (3.1)
below.
s(t) = - s(t-kN2 ). (3.1)
This means that the second half of the periodic shape of s(t) for kN2 t kN is the reflec-
tion of its first half about the origin. For the general shapes in B k , the first half of the
periodic shape, s(t) , can be any shape at all, provided k
N
ts2
0
2|)(| dt . This genera-
lization of Fourier analysis, called adaptive discrete Fourier analysis, is accomplished
via the periodic step functions f jnk ,, defined by Foutz and Lee ( 2000 ) instead of sinusoid
functions and via their properties as follows.
8/11/2019 Dissertation New
31/97
24
3.2 The periodic step functions f jnk ,,
When k > 0, the periodic step function sin( 2 t2 ) jumps to 0, 1, and 1 at the times t
where sin( t )=0, 1, and 1 respectively. It follows that the complex valued step function
exp(i2
t2) = cos(
2
t2) + i*sin(
2
t2) (3.2)
jumps to 1, i, -1 and i with frequency , See Figure 3.1 below. The function exp( i2
+ u
t
2) is a version of (3.2) that is shifted backwards by
2u time periods. For k > 0
and k = Nk2 , let particular time-shifted versions of (3.2) be denoted by
f jnk ,, (t) = exp(i2
+
+ 1
122
n
jtk
),
f jnk ,, (t) = exp(i2
+
n
jtk 122
),
f jn,,0 (t) = 1, for k = 0; where j = 1, 2, ,n. (3.3)
The functions f jnk ,, (t) and f jnk ,, (t) are complex conjugates. The appearance of some of
these functions is shown in Figure 3.2.
8/11/2019 Dissertation New
32/97
25
Figure 3.1. The periodic step function f jnk ,, in the complexplan
-1 -0.8 -0.6 -0.4 -0.2 0 0.2 0.4 0.6 0.8 1-1
-0.8
-0.6
-0.4
-0.2
0
0.2
0.4
0.6
0.8
1
The periodic step function in the complex plane
8/11/2019 Dissertation New
33/97
26
Figure 3.2. the step function f j,6,2 (t)
-3 -2 -1 0 1 2 3
-1
0
1
the s tep func t ion w i th k=2 , n= 6
j=3,im
ag
-3 -2 -1 0 1 2 3
-1
0
1
j=6,real
-3 -2 -1 0 1 2 3
-1
0
1
j=6,imag
-3 -2 -1 0 1 2 3
-1
0
1
the step function with k=2, n= 6
j=2,r
eal
-3 -2 -1 0 1 2 3
-1
0
1
j=2,imag
-3 -2 -1 0 1 2 3
-1
0
1
j=3,real
8/11/2019 Dissertation New
34/97
8/11/2019 Dissertation New
35/97
28
The proof of the Theorem 4.1 is given in Appendix 1. This theorem is the first and most
important result in this research. This is due to the fact that the following Corollary 3.1,
and the property of s mn, are based on this theorem, and the adaptive Fourier analysis of x
proceeds by investigating the properties of s mn, (t).
3.2.5. The decomposition of L 2 .
Corollary 3.1. L 2 has another decomposition as
L 2 = + B 1 + B 0 + B1 + ,
where B k is the subspace of L2 that contains all periodic functions that have the
form s ][k (t), in which s ][k (t) = - s ][k (t-k ), for k 0; and B 0 ={c, c is complex-valued
scale}.
The proof of the Corollary3.1 is also given in Appendix 1.
3.4. Multiresolution analysis
The sequence of spaces { L nKB ,, } ZK represents a ladder of subspaces of increasing
resolution as K increases, and it has the following properties:
1. L nB ,1, L nB ,2, ;
2. ( hZK LnKB ,, )= L nB ,0, = { 1 };
3.n
lim( ZK
L nKB ,, ) = L2 .
8/11/2019 Dissertation New
36/97
8/11/2019 Dissertation New
37/97
30
3.5.2. The digital data set, x = (x(t1 ), , x(tN )) is replaced by a continuous-time function
x e (t ) in B 1 +B 0 +B1 . It is required that x e (t j ) = x (t j ), at j = 1, 2,, N.
Theorem 3.2 For digital data set x = ( x(t1
), x(t2
), , x(tN ) ) , where t1
, , tN are not
necessarily equally-spaced and tN = N, there always exists a step function x e (t) L 1,B =
B 1 +B 0 +B1 such that
x e (t j ) = x (t j ) at j = 1, 2,, N.
The proof of this theorem is in Appendix 2.
3.5.3 x e (t ) is projected onto L nKB ,, by using a weight inner product m,3 .
3.5.3.1. Construct a weight inner product
The weight function
w m (t) =
+
+
+
otherwise
tttifm
tttifm
tttifm
m
mNmN
mm
mm
,
,
,
,
1
21
21
21
221
2
21
121
1
(3.6)
steps up to the value m near t = t 1 , , t N and approaches 0 elsewhere. w m (t) approaches
a Dirac -function, (t), as m . It is used to construct the inner product
m,3 =N
1
N
m dttztytw0
* )()()(
with the properties:
mlim
N
1dttztytw
N
m 0
2|)()(|)( =N
1 2
1
|)()(|=
N
j
jj tzty , (3.7)
8/11/2019 Dissertation New
38/97
31
andm
lim m,3 =N
1
=
N
j
jty1
2 )( , (3.8)
whenever y and z are in L nKB ,, .
3.5.3.2 The projection
The inner product m,3 is used to project x e (t) onto L nKB ,, for any K 1. The projec-
tion, s mn, = P m,3 ( x e | L nKB ,, ), has the property below
x(t) =n
limm
lim s mn, (t) (3.9)
at each t = t1 , , tN . This is proven in Appendix 3. The adaptive Fourier analysis of
x proceeds by investigating the properties of s mn, (t) for increasing n and for large m.
3.5.4 Adaptive Fourier series representation and ANOVA
The spaces B nk, are not orthogonal with respect to the usual inner products < y,z>1 , 2 and < y, z > m,3 . However, the spaces are orthogonal with respect to a nonstandard
inner product 4 that is defined in Appendix 4. It follows that s mn, has an adaptive
Fourier series representation
s mn, (t) = =
K
Kk
mnks ,, (t), (3.10)
where s mnk ,, (t) = P 4 ( s mn, | B nk, ) is the projection of s mn, onto B nk, with respect to 4 . The corresponding adaptive Fourier series representation for x is
x(t) =n
limm
lim =
K
Kk
mnks ,, (t), (3.11)
8/11/2019 Dissertation New
39/97
8/11/2019 Dissertation New
40/97
8/11/2019 Dissertation New
41/97
34
s mn, (t) = =
D
r
rr twd1
)( , (3.15)
where dr= < x e , w r> m,3 . (3.16)
Let a sr, be the element of MD in the rth row and s th column, and let a*,sr be its complex
conjugate. Substituting MD v for w in (3.15) and (3.16) gives
s mn, (t) = = =
D
r
ssrr
r
s
tvad1
,1
)( = = =
D
s
ssrr
D
sr
tvad1
, )( , (3.17)
and dr= mr
s
sesr vxa ,31
*, , > 4 , it follows from ( 3.17 ) that the projection
P 4 ( s mn, | B nk, ) is given by
s mnk ,, (t) = P 4 ( s mn, | B nk, ) =
++
++= =
nnKk
nKksssrr
D
rtvad
)(
1)(,
1 )( . (3.19)
Definition. Fix a positive integer n. The adaptive discrete Fourier transform (ADFT ) of
the continuous-time extension x e (t) of the digital time series x is the set of n N scalars
jnk ,, = < x e , f jnk ,, > m,3
for j = 1, 2, , n and K k K.
Property. ADFT jnk ,, = < x e , f jnk ,, > m,3 are relative robust to the construction of x e (t),
they only require that x e (t i ) = x(t i ) for i = 1, 2, ... , N.
8/11/2019 Dissertation New
42/97
8/11/2019 Dissertation New
43/97
36
4. Examples
Example 4.1 The time series data set x(1), x(2), , x(72) in Table 4.1 and Figure 4.1 is
taken from Brockwell and Davis (1991, p7), It contains the monthly accidental deaths in
the U.S.A. from January, 1973 through December, 1978.
Table 4.1Monthly Accidental Deaths in the U.S.A., 1973-1978
1973 1974 1975 1976 1977 1978Jan. 9007 7750 8162 7717 7792 7836Feb. 8106 6981 7306 7461 6957 6892Mar. 8928 8038 8124 7776 7726 7791
Apr. 9137 8422 7870 7925 8106 8129May 10017 8714 9387 8634 8890 9115Jun. 10826 9512 9556 8945 9299 9434Jul. 11317 10120 10093 10078 10625 10484Aug. 10744 9823 9620 9179 9302 9827Sep. 9713 8743 8285 8037 8314 9110Oct. 9938 9129 8433 8488 8850 9070Nov. 9161 8710 8160 7874 8265 8633Dec. 8927 8680 8034 8647 8796 9240
0 1 0 2 0 3 0 4 0 5 0 6 0 7 0 8 06 5 0 0
7 0 0 0
7 5 0 0
8 0 0 0
8 5 0 0
9 0 0 0
9 5 0 0
1 0 0 0 0
1 0 5 0 0
1 1 0 0 0
1 1 5 0 0
F i g u re 4 . 1 M o n t h ly a c c i d e n t a l d e a th s i n U . S . A . , 1 9 7 3 -1 9 7 8
8/11/2019 Dissertation New
44/97
8/11/2019 Dissertation New
45/97
38
Table 4.2 *10^5
K Period
kN
ANOVA
|P 2 (x|A k )|22 +|P 2 (x|A k )|
22
Adaptive
|s mnk ,, |24 +|s mnk ,, |
24
k=1 72 1.2080 1.0852
k=2 36 0.1747 0.3414k=3 24 0.3197 0.1907k=4 18 0.1009 0.6844k=5 72/5 0.0492 0.1493k=6 12 5.2262 6.0817
Total 7.0787 8.5327
Figure 4.2 Frequency component with k=6 in Example 4.1
8/11/2019 Dissertation New
46/97
39
1 1.5 2 2.5 3 3.5 4 4.5 5 5.5 60
2
4
6
8x 10
5 Figure 4.3: Adaptive Fourier line spect rum for Exam ple 4.1
1 1.5 2 2.5 3 3.5 4 4.5 5 5.5 60
2
4
6x 10
5 Figure 4.4 Fourier line spectrum in Example 4.1
(2). To illustrate that adaptive Fourier analysis can be applied to unequally-spaced time
series data, we delete the Marchs records, then the data is reduced to an unequally-spa-
ced time series with 66 sample points, x = ( x(t1 ), x(t 2 ), ..., x(t 66 )). In this case, classical
Fourier ANOVA cant be used. However, adaptive Fourier ANOVA can be used to re-
veal the frequency component shapes and discovering periodicity. Similar to the above,
we pick K = 6, n =2 and m=1000 respectively. The resulting adaptive Fourier ANOVA is
given in Table 4.3, it also reveals that this unequally-spaced time series has the same
major frequency component with k = 6. The time series x = ( x(t1 ), x(t 2 ), ... , x(t 66 ) ) is
plotted on Figure 4.5 and the major frequency component s mn,,6 + s mn,,6 is plotted on
Figure 4.6.
8/11/2019 Dissertation New
47/97
40
Table 4.3 *10^5
K Period
kN
Adaptive ANOVA
|s mnk ,, |24 +|s mnk ,, |
24
k=1 72 1.1436
k=2 36 0.7401k=3 24 0.1260k=4 18 0.4762k=5 72/5 0.1795k=6 12 6.5532
Total 9.2187
(3). If we take Januarys record out from the first year and Februarys record out from
the second year, , Junes record out from the last year, then the data will be reduced to
an another unequally-spaced time series with 66 sample points, x = ( x( t1 ), x( t 2 ), ...
8/11/2019 Dissertation New
48/97
41
, x(t 66 ) ). In this case, adaptive Fourier ANOVA can be also used to reveal the frequency
component shapes and discovering periodicity. Similar to the above, we pick K= 6, n = 2
and m =1000 respectively. The resulting Adaptive Fourier ANOVA is given in Table 4.4.
It also reveals that this unequally-spaced time series has the same major frequency com-
ponent with k=6. The time series x = ( x(t 1 ), x(t 2 ), ..., x(t 66 )) is plotted on Figure 4.7 and
the major frequency component s mn,,6 + s mn,,6 is plotted on Figure 4.8.
Table 4.4 *10^5
K Periodk
N Adaptive ANOVA|s mnk ,, |24 +|s mnk ,, |
24
k=1 72 1.1122k=2 36 0.5772k=3 24 0.1494k=4 18 0.5806k=5 72/5 0.1614k=6 12 5.8413
Total 8.4222
8/11/2019 Dissertation New
49/97
42
(4). If we take randomly 10 records out from the original data set, then the data will be
reduced to an unequally-spaced time series with 62 sample points, x = ( x(t1 ), x(t 2 ), ... ,
x( t 62 ) ). The resulting adaptive Fourier ANOVA is given in Table 4.5, it also reveals
that this unequally-spaced time series has the same major frequency component with k=6.
The time series x = ( x( t1 ), x( t 2 ), ... , x(t 62 ) ) is plotted on Figure 4.9 and the major fre-
quency component s mn,,6 + s mn,,6 is plotted on Figure 4.10.
8/11/2019 Dissertation New
50/97
43
Table 4.5 *10^5K Period
kN
Adaptive ANOVA
|s mnk ,, |24 +|s mnk ,, |
24
k=1 72 1.3207k=2 36 0.3167
k=3 24 0.1788k=4 18 0.6656k=5 72/5 0.4468k=6 12 6.5826
Total 9.5111
(5). We repeat (4) above again, and obtain similar results shown in Table 4.6 and Figure
4.11 and 4.12.
8/11/2019 Dissertation New
51/97
44
Table 4.6 *10^5K Period
kN
Adaptive ANOVA
|s mnk ,, |24 +|s mnk ,, |
24
k=1 72 1.1929k=2 36 0.4377
k=3 24 0.1794k=4 18 0.5834k=5 72/5 0.1726k=6 12 6.8078
Total 9.3739
This example implies that when a small fraction of the original data, 10/72 14% in this
example, is removed or missed, the adaptive Fourier analysis can still detect the main
frequency components in the original time series data.
8/11/2019 Dissertation New
52/97
45
Example 4.2. To illustrate that for some signals with jumps, adaptive Fourier analysis
outperforms Fourier analysis, we take an equally-spaced sample s = (s1, s2, , s192 ) from
the signal s(t) with sample size N =192 to do adaptive Fourier analysis and Fourier analy-
sis. Here s(t) is given by
s(t) =
8/11/2019 Dissertation New
53/97
46
component with k = 2 and thus in Fourier ANOVA, 20.5% of the sample energy has
leaked onto a wide frequency range, which implies that in this case adaptive Fourier
analysis outperforms Fourier analysis.
Table 4.2(1)
K Period
kN
ANOVA
|P 2 (x|A k )|22 +|P 2 (x|A k )|
22
Adaptive ANOVA
|s mnk ,, |24 +|s mnk ,, |
24
k=1 192 0.0000 0.0008
k=2 96 2.2004 2.7639k=3 62 0.0000 0.0008
Total 2.2004 2.7647
8/11/2019 Dissertation New
54/97
47
0 1 2 3 4 5 6-2.5
-2
-1.5
-1
-0.5
0
0.5
1
1.5
2
2.5
Figure 4.2(3): k=2 freq. component(-) in Fourier Analysis & original signal(.)
Fourier freq. comp.(k=2)
originial signal
To illustrate that adaptive Fourier analysis is different from wavelet analysis, we apply
wavelet analysis to the sample data s = ( s1, s2, , s192 ) in Figure 4.2(1). The multiresolu-
tion decomposition of the signal is plotted in Figure 4.2(4) and the time-scale plot for the
signal is plotted on Figure 4.2(5). In wavelet analysis, fine scale wavelet functions usu-
all reveal where or when the signal has high frequency components in a loose sense.
However, in this case, all the fine scale wavelet functions focus on the discontinuous
points of the signal shown in Figure 4.2(4) andespecially in Figure 4.2(5). Therefore, in
this case, wavelet analysis cant provide any information about the frequency component.
8/11/2019 Dissertation New
55/97
48
S5
D5
D4
D3
D2
D1
Data
0 50 100 150
Figure 4.2(4): Multiresolution decomposition of the signal
Figure 4.2(5). Time-scale plot for the signal
Time
1/Scale
0 50 100 150
0.0
0.2
0.4
0.6
0.8
1.0
8/11/2019 Dissertation New
56/97
49
Example 4.3. To illustrate that for sinusoidal signal, with equally-spaced data, Fourier
analysis and adaptive Fourier analysis are almost the same, we take an equally-spaced
sample s = ( s 1 , s 2 , ..., s 60 ) from the signal s( t ) = 10*cos(2 t/15) + 3cos(2 t/10) over
[1, 30], sample size N = 60. The sample series, s = ( s 1 , s 2 , ... , s 60 ) is plotted on Figure
4.3(1). The Fourier ANOVA in Table 4.3(1) reveals that the sample series contains a
major component k=2 and a minor one k = 3. In adaptive Fourier ANOVA, the frequency
number K = 3, the dimensions n = 5 and m = 500 are used, and the ANOVA result is also
given in Table 4.3(1), again showing that the adaptive frequency component for k = 2,
namely s mn,,2 + s mn,,2 , is the major component ; for k = 3, s mn,,3 + s mn,,3 is the minor
component. The power distribution for the first three frequencies of adaptive Fourier
analysis is plotted on Figure 4.3(3); the one for Fourier Analysis is plotted on Figure
4.3(4); s mn,,3 + s mn,,2 + s mn ,,1 + s mn,,1 + s mn,,2 + s mn ,,3 is plotted on Figure 4.3(2); s mn,,3
+s mn,,3 is plotted on Figure 4.3(5); and s mn,,2 + s mn,,2 is plotted on Figure 4.3(6). Table
4.3(1) shows that Fourier and adaptive Fourier ANOVA are very close; Figure 4.3(5) and
Figure 4.3(6) show that major and minor adaptive frequency components both have a
sinusoidal shape. The results together show that for equally-spaced sampled series s =
(s1 , s 2 , ..., s 60 ) from a sinusoidal signal, Fourier ANOVA and adaptive ANOVA are very
close.
8/11/2019 Dissertation New
57/97
50
Table 4.3(1)
K period
kN
ANOVA
|P 2 (x|A k )|22 +|P 2 (x|A k )|
22
Adaptive ANOVA
|s mnk ,, |24 +|s mnk ,, |
24
k=1 60 0.162 0.1744k=2 30 47.206 46.8258k=3 20 5.2238 5.2992Total 52.5918 52.2994
8/11/2019 Dissertation New
58/97
51
1 1.2 1.4 1.6 1.8 2 2.2 2.4 2.6 2.8 30
10
20
30
40
50
Figure 4.3(4): Fourier line s pectrum
1 1.2 1.4 1.6 1.8 2 2.2 2.4 2.6 2.8 30
10
20
30
40
50
Figure 4.3(3): A daptive Fourier line spect rum
8/11/2019 Dissertation New
59/97
52
Example 4.4. To illustrate that adaptive Fourier analysis can be applied to unequally-
spaced data from a sinusoidal signal, we use the uniform distribution to take a random
sample s = ( s(t 1 ), s(t 2 ), ..., s(t 60 )) from the signal s(t) in Example 4.3 above. The sample
size is N = 60. The sample series s = ( s(t 1 ), s(t 2 ), ... , s(t 60 ) ) is plotted on Figure 4.4(1).
Similarly to Example 4.3, we pick K = 3, n = 7 and m = 500 respectively. The resulting
adaptive Fourier ANOVA is given in Table 4.4(1), and plotted on Figure 4.4(1); s mn,,3 +
s mn,,2 + s mn ,,1 + s mn,,1 + s mn,,2 + s mn,,3 is plotted on Figure 4.4(2); s mn,,2 + s mn,,2 is plotted
on Figure 4.4(3); and s mn,,3 + s mn,,3 is plotted on Figure 4.4(4). Table 4.4(1) shows that
adaptive Fourier analysis can reveal the two frequency components with k = 2 and k = 3.
Figure 4.4(4) and Figure 4.4(5) show that these two frequency components are sinu-
soidal .
Table 4.4(1)
K period
kN
Adaptive ANOVA
|s mnk ,, |24 +|s mnk ,, |
24
k=1 60 0.1624k=2 30 49.4853k=3 20 3.5624Total 53.2102
8/11/2019 Dissertation New
60/97
8/11/2019 Dissertation New
61/97
54
Example 4.4 implies that for an unequally-spaced time series where the observation times
follow uniform distribution , the adaptive Fourier analysis can still detect the main fre-
quency components in the original time series data.
Example 4.5. To illustrate that for non-sinusoidal signal, adaptive Fourier analysis out-
performs Fourier analysis, we take an equally-spaced sample s = ( s1 , s 2 , ..., s 60 ) from
the signal s(t) with sample size N = 60 to do adaptive Fourier analysis. The signal s(t) is
given by
8/11/2019 Dissertation New
62/97
55
s(t) =
8/11/2019 Dissertation New
63/97
56
8/11/2019 Dissertation New
64/97
57
1 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8 1.9 20
1
2
3
Figure 4.5(3): A daptive Fourier line spect rum
1 1.2 1.4 1.6 1.8 2 2.2 2.4 2.6 2.8 30
0.5
1
1.5
2
Figure 4.5(4): Fourier line spec trum
8/11/2019 Dissertation New
65/97
58
5. Conclusion and Future Research
5.1 Summary and Conclusion
The goal of this dissertation is to develop theory and methods that can be applied to
equally and unequally-spaced time series in which the frequency components of time
series may take general periodic shapes that include sinusoids as special cases. The re-
sults of the research are summarized as follows:
(1) Theorem 3.1 Any function s(t) in L 2 that satisfies s(t) = -s(t-k ) is the limit of
step functions s n (t) in B nk, as n , in the L2 sense.
(2) Corollary 3.1. L 2 has another decomposition as
L 2 = + B 1 + B 0 + B1 +
where B k is the subspace of L2 that contains all periodic functions that have
the form s ][k (t), in which s ][k (t) = - s ][k ( t-k ), for k 0; and B 0 = {c, c is
complex-valued scale}.
(3) Through a weight inner product, < y, z > m,3 , a new method of projection has
been developed, which can be applied to project x e (t) onto L nKB ,, as s mn, (t) =
P m,3 ( x e (t) | L nKB ,, ), where x e (t) is the continuous- time extension of equally
or unequally-spaced time series. The adaptive Fourier analysis of time series
x proceeds by investigating the properties of the projection, s mn, (t).
8/11/2019 Dissertation New
66/97
59
(4) By theorem 3.1 an important property of s mn, (t) can be proved as
x(t) =n
limm
lim s mn, (t),
at each t = t1 , , tN .
(5) A multiresolution analysis (MRA) of L 2 has been presented, which showed
that the step functions used in this research can generate a MRA of L 2 .
(6) Through a nonstandard inner product 4 , adaptive Fourier ANOVA has
been developed, which can be applied to equally and unequally-spaced time
series.
(7) Examples: Example 4.1 gives an application of adaptive Fourier analysis to a
real data set. Example 4.2 illustrates that for some signals with jumps, adap-
tive Fourier analysis outperforms Fourier analysis and is different from wave-
let analysis. Example 4.3 illustrates that adaptive Fourier analysis can be
applied to the time series with general patterns that include sinusoids as
special cases. Example 4.4 illustrates that adaptive Fourier analysis can direct-
ly deal with unequally-spaced time series with sinusoidal frequency compo-
nents. Example 4.5 illustrates that adaptive Fourier analysis outperforms
Fourier analysis for a non-sinusoidal signal.
5.2 Proposed Future Research
(1) A better algorithm is needed for computing the adaptive Fourier trans-
form. Before a better algorithm is developed, use of adaptive Fourier
8/11/2019 Dissertation New
67/97
60
analysis in analyzing time series with large sample size would be time-
consuming.
(2) To do statistics tests for hidden periodic components, relevant theory
and methods for adaptive Fourier analysis are needed.
(3) The orthogonal series approach to nonparametric regression has be-
come popular lately. People can approximate a function by polynomi-
als, sinusoids, step functions, and wavelets and apply these approxima-
tions to relevant nonparametric regressions. Nonparametric regression
by the step functions used in this research seems promising and needs
to be developed. We expect that this kind of nonparametric regression
would be applied to unequally-spaced data observations and would
have fewer terms in the regression models for some cases, i.e. would
use fewer degrees of freedom.
8/11/2019 Dissertation New
68/97
61
Bibliography
Anderson, J.A. (1969),Real Analysis. Gordon and Breach Science Publishers, New
York.
Bruce, Andrew and Gao, Hong-Ye (1996),Applied Wavelet Analysis With S-Plus.Springer-Verlag Inc, New York.
Barthes, D. (1995), Cleaning Power Spectra of Short, Unevenly Spaced Time-SeriesApplication to LPVS,Astronomy & Astrophysics Supplement Series. 111(2):373-386 Jun 1995.
Beauchamp, K.G. (1984),Applications of Walsh and Related Functions. London:Academic Press.
Brockwell, P.J. and Davis, R.A. (1990), Time Series: Theory and Methods(SecondEdition). Springer-Verlag Inc, New York.
Bowen, Ray M. and Wang, C. C. (1976),Introduction to Vectors and Tensors Linear andMultilinear Algebra. Plenum Press, New York.
Baisch, S. and Bokelman, GHR (1999), Spectral Analysis with Incomplete Time Series:An Example from Seismology, Computers & Geosciences. 25(7): 739-750 Aug.1999.
Carothers, N.L. (2000),Real Analysis,Cambridge University Press.
Dorny, C. Nelson (1980),A Vector Space Approach to Models and Optimization. RobertE. Krieger Publishing Company.
Edelson, R.A. and Krolik, J.H. (1988), The Discrete Correlation Function: A NewMethod For Analyzing Unevenly Sampled Variability Data, The AstrophysicalJ., 333: 646-659.
Engle, RF and Russell JR (1998), Autoregressive Conditional Duration: A New Modelfor Irregularly Spaced Transaction Data,Econometrica.66(5): 1127-1162.
Foutz, R.V. and Lee, H. (2000), Adaptive Fourier Series and The Analysis OfPeriodicities in Time Series Data, J. Time Ser. Anal. Vol. 21, No. 6 p649-662.
Frazier, Michael W. (1999),An Introduction to Wavelets Through Linear Algebra,Springer-Verlag Inc, New York.
8/11/2019 Dissertation New
69/97
62
Golubov, B., Efimov, A. and Skvorstsov, V. (1991), Walsh Series and Transforms:Theory and Applications, Kluwer Academic Publishers.
Good, I. J. and Doog, K. C. (1958), A Paradox Concerning Rate of Information,Information and Control, Volume 1, 113-126.
Halmos, P. R. (1957),Introduction to Hilbert Space and the Theory of SpectralMultiplicity. New York: Chelsea.
Koopmans, L.H. (1974), The Spectral Analysis of Time Series. New York: Academic Press.
Kuhn, J.R. (1982), Recovering Spectral Information From Unevenly Sampled Data:Two Machine-efficient Solutions, The Astronomical J.,Vol. 87, No. 1, pp 196-202.
Lomb, N.R. (1976), Least-Squares Frequency Analysis of Unequally Spaced Data,Astrophysics and Space ScienceVol. 39, pp447-462.
Marvasti, Farokh (2001),Nonuniform Sampling Theory and Practice. KluwerAcademic/Plenum Publishers.
Meisel, D.D. (1978), Fourier Transforms of Data Sampled at Unequal Observational Intervals, The Astronomical J., Vol. 83, No. 5, P538-545.
Priestley, M.B. (1996), Wavelets and Time-dependent Spectral Analysis, J. Time Ser. Anal. 17, 85-103.
Rao, T. Subba, Priestley, M.B. and Lessi, O. (1997),Applications of Time SeriesAnalysis in Astronomy and Meteorology. Chapman & Hall.
Ruf T. (1999), The Lomb-Scargle Periodogram in Biological Rhythm Research:Analysis of Incomplete and Unequally Spaced Time-Series,Biological RhythmResearch30(2): 178-201 APR 1999.
Scargle, J.D. (1982), Studies in Astonomical Time Series Analysis. II. StatisticalAspects of Spectral Analysis of Unevenly Spaced Data, The Astronomical J.,263: 835-853.
Scargle, J.D. (1989), Studies in Astronomical Time Series Analysis. III. FourierTransforms, Autocorrelation Functions, and Cross-Correlation Functions ofUnevenly Spaced Data, The Astronomical J., Vol. 343, pp874-877.
Schimmel, M. (2001), Emphasizing Difficulties in The Detection of Rhythms withLomb-Scargle Periodograms,Biological Rhythm Research.32(3): 341-345 Jul.2001.
8/11/2019 Dissertation New
70/97
63
Schulz, M. and Stattegger, K. (1997), Spectrum: Spectral Analysis of Unevenly Spaced
Paleoclimatic Time Series, Computers & Geosciences23(9): 929-945. 11/1997.
Stoffer, D.S. (1991), Walsh-Fourier Analysis and Its Statistical Applications, J. Am.Stat. Assoc. 86, 461-85.
Swan, P.R. (1982), Discrete Fourier Transforms of Nonuniformly Spaced Data, TheAstronomical J.,Vol.87, No.11 p1608-1615.
Van, Dugen HPA, Olofsen, E. VanHartevelt, JH, Kruyt, EW (1999), A Procedure ofMultiple Period Searching in Unequally Spaced Time-Series with The Lomb-Scargel Method,Biological Rhythm Research, 30(2): 149-177 APR 1999.
8/11/2019 Dissertation New
71/97
64
Appendix 1
Theorem 3.1. Any function s(t) in L 2 that satisfies s(t) = - s(t -k ) for all t R , is the
limit of step functions s n (t) in B nk, as n , in the L 2 sense.
Proof: Let k be a positive integer, and n = 2 m for some positive integer m, G j and H j are
any real numbers, for j = 1, 2, , n. The function
x n (t) = )(
1
=
+n
j
jj HiG f jnk ,, (t) is in B nk, ,
since each f jnk ,, (t) is in B nk, . We pick the constants G j and H j symmetrically such that
G1 = - H1 ,
(G 2 , H 2 ) = (G n , - H n ),
(G 3 , H 3 ) = (G 1n , - H 1n ),
(G2n , H
2n ) = (G 2
2+n
, - H2
2+n
),
H1
2+n
= 0.
We denote the imaginary part of a function x(t) as I m {x(t)} and divide the interval - k1
t < -k1 +
k2 into n equally spaced small intervals as I1 = [- k
1 , -k1 +
kn2 ), I 2 = [- k
1 +kn2 , -
k1 +
kn22 ), , I l = [- k
1 +kn
l
2
)1( ,-k1 +
knl2 ), , and I n = [- k
1 +kn
n
2
)1( , -k1 +
knn2 ). Now note that
8/11/2019 Dissertation New
72/97
8/11/2019 Dissertation New
73/97
66
3 0 . When t I 3 = [ - k1 +
kn22 , -
k1 +
kn23 ),
I m {x n (t)} = I m { )(3
1
=
+j
jj HiG f jnk ,, (t)} + I m { )(4
=
+n
j
jj HiG f jnk ,, (t)}
= (G1 + G 2 +G 3 )+ (H 4 + + H n )
= G1 + G 2 +G 3 - H 2 - H 3
= s 2 + G 3 - H 3 ,
Let s 3 = s 2 + G 3 - H 3 , then
I m {x n (t)} = s 3 = s 2 + G 3 - H 3 , = G1 , for t I 3 = [ - k1 +
kn22 , -
k1 +
kn23 ).
4 0 . Similarly, for l 2n , we can obtain
I m {x n (t)} = s l
= s 1l +G l - H l for t I l .
and s l = (G1 ++ G l ) (H 2 ++ H l ) for 2 l 2n .
5 0 . For l=2
n +1, and t I12+
n,
I m {x n (t)} = I m { )(1
1
2
+
=
+
n
j
jj HiG f jnk ,, (t)} + I m { )(2
2
+=
+n
j
jjn
HiG f jnk ,, (t)}
= (G1 ++ G 12
+n)+ (H
22
+n++ H n )
= s2n + G
12+n
.
6 0 . Similarly, for l= 2n +2 and t I 22+n
, we have
I m {x n (t)} = s 22
+n
= s1
2+n
+G2n +H
2n .
8/11/2019 Dissertation New
74/97
67
7 0 For 2n +2 l n and t I l , similarly we have
I m {x n (t)} = s l
= s1
l +G2+
ln +H2+
ln .
Based on 1 0 through 7 0 above, we have
I m {x n (t)} =
++
++
.,
,
,
,
,
,
22
11
33
22
11
22
22
nn Itfors
Itfors
Itfors
Itfors
Itfors
Itfors
nn
nn
(A1)
where
++=
++=+=
+=
+=
+=
=
++++
++
+
.221
1212
11
3323
2212
11
222222
22
222
HGss
HGsHGss
Gss
HGss
HGss
Gs
nn
nnnnn
nn
nnn
(A2)
The equation (A2) above can be represented by matrix form as
8/11/2019 Dissertation New
75/97
8/11/2019 Dissertation New
76/97
69
Set s n = 1 and s j = 0 if 1 j < n, then the equation (A3 ) has an unique solution since
rank( C nn ) = n. Corresponding to the solution, I m { x n ( t ) }, is denoted as I m x nn, ( t ),
where
I m x ,n n ( t ) =
==
=
=
+
++
+
++
)...(,0
,1,1
)...(,0
,1
,1
)...(,0
423
1313
33
132
11
11
nn
nn
nn
nn
nn
nn
n
IItfor
ItforsItfors
IItfor
Itfors
Itfors
IItfor
,
For each l , let the step function I m x nln , ( t ) = I m x nn, ( t + knl2 ) be a location shifted
version of I m x nn, ( t ), then we obtain 2n-1 step functions, I m x nln , ( t ) in B nk, , for l=
0, 1, 2, , (n-1).
For any > 0, there is a step function H(t) with || I m {s(t)}IA (t)-H(t)|| 2 < k8 , where I A (t)
= 1 if t A = [-k1 , -
k1 +
k ) and I A (t) = 0 if t A = [- k
1 , -k1 +
k ), See Carothers (2000, p
350 ). If we define H(t) = - H( t-k ) for t [-
k1 +
k , -
k1 +
k2 ), and let H(t) repeat itself
periodically over interval of lengthk2 , and let A j = [ - k
1 +k
j )1( , -k1 +
k
j ), for j = 1, 2, ,
, 2k. Then, over C[0, 2 ], we have
8/11/2019 Dissertation New
77/97
70
|| I m {s(t)}-H(t)||22 =
=
k
j
2
1
|| IjA
(t) (I m {s(t)}-H(t)) ||22
< 2k(k8
) 2
= 42 , (A4)
where IjA(t) = 1 if t A j and I jA (t) = 0 if t A j .
Furthermore, since H( t ) is a step function, H( t ) must be constant on each of the open
intervals (t j , t 1+j ), where - k1 = t 0 < t1 < < t L = - k
1 +k2 , and j = 0, 1, 2, , L-1. Note
that step function is bounded, we can set M =),[ 211
maxkkk
t +| H(t)|. Considering the subinter-
vals I1 I 2 , I 3 I 4 , , I 14 n I n4 , where I1 I 2 I 3 I 4 I 14 n I n4 = [- k1 ,
-k1 +
k2 ), we define 2n constant values c1 , c 2 , , c n2 by
c i =
=
LjjsomeforIItandIItifM
LjforIItandIItiftH
iijii
iijii
0,,,
1...,,2,1,0,),(
212212
212212 (A5)
for i = 1, 2, , 2n.
Set h(t) = =
n
l
nlml txIc1
),12( )( , (A6)
8/11/2019 Dissertation New
78/97
71
then, h(t) is a step function in B nk, , and satisfies that h(t)= -h( t- k ) and |h(t)-H(t)| 2M.
Thus, over C[k1 , -
k1 +
k2 ), we have
|| h(t)-H(t)|| 22 L(2M)2 *
nk
=
nk
LM 24. (A7)
When n > 2216
LM , based on (A7), over C[0, 2 ], we obtain
|| h(t)-H(t)|| 22 42 . (A8)
(A4) and (A8) together show
|| I m {s(t)}-h(t)|| 2 || I m {s(t)}-H(t)|| 2 + || H(t)-h(t)|| 2
2 + 2
=
Similarly, the real part of s(t) has the same conclusion , therefore conclusions together
prove the Theorem 3.1.
Corollary 3.1 L 2 has another decomposition as below
L 2 = + S 1 +S 0 + S1 +
8/11/2019 Dissertation New
79/97
8/11/2019 Dissertation New
80/97
73
Appendix 2
Construct x e (t) as follows.
Case 1. Assume there are not t i = t j 2N for any i , j {1, 2, , N} where t N = N. Set
t =Njj ',1
min {|t j -(t 'j 2N )|. Define x e (t) as below
.
x e (t) =
+
8/11/2019 Dissertation New
81/97
74
x e (t) =
+
8/11/2019 Dissertation New
82/97
75
Appendix 3
We need to prove
x(t) =n
limm
lim s mn, (t) , for t = t 1 , , tN . (3.9)
Note the construction of x e (t) in Appendix 2, we know that x e (t) is continuous at all t =
t j , j = 1, 2, , N. By Theorem 3.1, x e (t) is a limit of step functions e n,1 (t) in L nB ,1, for
increasing n. In addition, x e (t) is continuous at t = t 1 , , t N , based on its construction in
Appendix 2. Thus, given any
> 0, when n is large enough, we have
| x e (t) - e n,1 (t) | < 21
, for t = t1 , , tN . (A3.1)
which implies that
nlim
Nttt 1
sup | x e (t) - e n,1 (t) | = 0. (A3.2)
Now note x(t) = x e (t) , for t = t 1 , , tN . It follows from (A3.1) or (A3.2) that
nlim | x(t) - e n,1 (t) | = 0 , for t = t 1 , , tN . (A3.3)
The function e n,1 (t) is in L nKB ,, when K 1, and the projection s mn, (t) = P m,3 ( x e |
L nKB ,, ) minimizes | x e (t)- s mn, (t) | m,3 over functions in L nKB ,, . Thus,
| x e - s mn, | m,3 | x e (t) - e n,1 (t) | m,3 . (A3.4)
It follows from (A3.4) that
N
1dttstxtw mne
N
m
2,
0|)()(|)( N
1dttetxtw ne
N
m
2,1
0|)()(|)( . (A3.5)
Taking limits on both sides of (A3.5) for increasing m, we have
8/11/2019 Dissertation New
83/97
76
m
limN
1dttstxtw mne
N
m
2,
0|)()(|)( mlim N
1dttetxtw ne
N
m
2,1
0|)()(|)( . (A3.6)
By (3.7) in section 3.5, (A3.6) becomes
N1
=
N
j 1
| x e (t j )- s mn, (t j )|2
N1
=
N
j 1
| x e (t j )- e n,1 (t j )|2 . (A3.7)
Because x(t) = x e (t) at t = t 1 , , tN , and from (A3.7), we obtain (A3.8) as below,
mlim
N
1
=
N
j 1
| x (t j )- s mn, (t j )|2 =
mlim
=
N
j 1
| x e (t j )- s mn, (t j )|2
N
1
=
N
j 1
| x e (t j )- e n,1 (t j )|2
=N
1
=
N
j 1
| x (t j )- e n,1 (t j )|2 . (A3.8)
It follows from (A3.8) that
mlim
=
N
j 1
|x (t j )- s mn, (t j )|2
=
N
j 1
| x (t j )- e n,1 (t j )|2 . (A3.9)
By taking limits on both sides of (A3.9) for increasing n, and by (A3.3), we have
nlim
mlim
=
N
j 1
| x (t j )- s mn, (t j )|2
nlim
=
N
j 1
| x (t j )- e n,1 (t j )|2
= 0. (A3.10)
Conclusion (3.9) follows from (A3.10).
8/11/2019 Dissertation New
84/97
77
Appendix 4
A nonstandard inner product < y, z > 4 may be defined on L nKB ,, = B nK, + + B n,1 +
+ B n,1 + + B nK, as follows: Theorem 1 of Foutz and Lee (2000) shows that each
y in L nKB ,, has an unique representation,
y(t) = =
K
Kk
ky (t) , (A4.1)
where y k (t) is a unique function in B nk, for each k. Shift each frequency component in
y k (t) in (A4.1) by the fraction u of its period to obtain
y u (t) = =
K
Kk
ky (t + kuN ). (A4.2)
Define the inner product between y and z in L nKB ,, to be
< y, z > 4 = 1 du . (A4.3)
In Foutz and Lee (2000) it is proven that the spaces B nk, for k = -K, , K are orthogonal
with respect to < y, z > 4 .
Appendix 5.
It remains to verify (3.12). The left side of (3.12) is
8/11/2019 Dissertation New
85/97
78
| =
K
Kk
mnk ts )(,, |24 ,
This equals
=
K
Kk
mnk ts24,, |)(| ,
because the frequency component s mnk ,, ( t ) is in B nk, and because the spaces B nk, are
orthogonal with respect to the inner product < y, z > 4 . This sum in turn equals the right
side of (3.12), because s mnk ,, (t) is periodic with period kN and therefore
| s mnk ,, (t) | 24 = 1
0 1N2
0 ,, |)(| +N
mnkk
uNts dtdu
= 1
0
1
N
2
0,, |)(|
N
mnk ts dtdu
=N
1 20
,, |)(|N
mnk ts dt.
8/11/2019 Dissertation New
86/97
79
Appendix 6. Background
The concepts and results listed in this Appendix will provide an adequate background for
theoretical results and their proofs in Chapter 3 and the Appendix1-5. All conceptions and
their properties used in Chapter3 and Appendix1-5, such as vector space, direct sum, inner
product, Hilbert space, weight inner product, orthogonal projection, and Gram-Schmidt
procedure, are presented in this Appendix.
A6.1 Vector Spaces
Definition 6.1. (Vector space) Let F be a field, and V a set with operations addition and
scale multiplication. V is a vector space if it satisfies the following properties:
1) For operation addition +, V has properties (1) through (5) as below.
for anyx, y V, x +yV (closure for addition).
(1)(x + y) +z=x + (y + z) (Associativity for addition).
(2)x + y = y + x(commutativity for addition).
(3)There is an element 0 V such that x + 0 = x for anyx V (Existence of
additive identity).
(4)For anyx V, there is an element xsuch thatx + (-x) = 0.
2) For operation scale multiplication, V has properties 6 through 10 as below for any ,
F and anyx, yV.
(6) x V (Closure for scale multiplication).
(7) (x)= ()x (Associativity for scale multiplication).
(8) (+)x= x+ x (First distributive).
8/11/2019 Dissertation New
87/97
80
(9) (x + y) = x +y (Second distributive).
(10) 1x = x.
Remark. If the field F is real numbers R, the space V is called a real vector space;
similarly if the field F is complex numbers C, the space V is called a complex vector
space.
Definition 6.2. ( Linear independent ) A set of vectors { x 1, x 2 , , x n } in a vector
space V is said to be linearly dependent if there is a set of scales { 1 , 2 , , n } not all
zero, such that
1 x1 + 2 x 2 + + n x n = 0. (6.1)
If (6.1) holds only when i
= 0 for any i = 1, 2, , n, we say that {x
1, x
2, , x
n } are
linearly independent.
Definition 6.3. (Basis and dimension) Let {x 1, x 2, , x n } be linearly independent set
in vector space V, if for any vectorxV , {x1, x 2, , x n, x} are linearly dependent,
we say that { x 1 , x 2, , x n } is a basis of V and that V is finite dimensional and the
dimension of V is n, written dim V, namely n=dim V.
8/11/2019 Dissertation New
88/97
81
Definition 6.4.(Span)Let V be a vector space, and U V. The span of U is the set of all
linear combinations of elements of U.
Definition 6.5. ( Subspace ) A nonempty subset U of a vector space V over F is a sub-
space if:
x, y U x+ y U, for anyx, y U, and any , F.
Definition 6.6. (Sum of spaces)Let U and W be subspaces of a vector space V. The sum
of U and W, written U+W, is the set
U+W= {x + y|xU,yW}.
The intersection of U and W, written UW , is the set
UW = {x|xU andx W}.
The union of U and W, written UW, is the set
UW = {x|x U or x W }.
Some properties of these operations are stated in the following theorems. ( Theorem 6.1
through Theorem 6.4), see Bowen and Wang(1976, P53).
Theorem 6.1. If U and W are subspaces of V, then U+W is a subspace of V.
8/11/2019 Dissertation New
89/97
82
Theorem 6.2. If U and W are subspaces of V, then U and W are also subspaces of U+W.
Theorem 6.3. If U and W are subspaces of V, then the intersection UW is a subspace
of V.
Theorem 6.4. Let xbe a vector in U + W, where U and W are subspaces of V. The
decomposition ofx U+W into the formx = y + z, where y U andz W, is unique if
and only if UW = { 0 }.
Definition 6.7.(Direct sum)The sum of two subspaces U and W in V is called the direct
sum of U and W and written by
U W, if UW = { 0 }.
A6.2 Inner product spaces , Hilbert spaces and their properties
Definition 6.8. (Inner product space) A complex vector space V is said to be an inner
product space if there exists an operation called inner product by which any ordered pair
of vectorsyandzin V determines an element of C, written by such that:
(a) =_______
, >< yz , where_______
, >< yz is the complex conjugate of ;
(b) = + ;
(c) = < y, z>;
8/11/2019 Dissertation New
90/97
83
(d) 0 and = 0 if and only if y= 0;
for allx, y, zV and C.
Definition 6.9. (Norm).The norm of an elementxof an inner product space is an opera-
tion, written by || ||, defined by
||x|| = >< xx, .
Theorem 6.5(The Schwarz inequality)
| | ||y || ||z||
is valid for any two vectorsy, zin an inner product space.
For Theorem 6.5 and its proof, see Bowen and Wang (1976, p60).
Theorem 6.6( The triangle inequality)
||y + z|| ||y || + ||z||
is valid for any vectors y, z in an inner product space.
For Theorem 6.6 and its proof, see Bowen and Wang (1976, p60).
Proposition 6.1. (Properties of the norm). A norm on an inner product space V over C
satisfies the following properties:
(a) ||y || = | | ||y ||;
8/11/2019 Dissertation New
91/97
8/11/2019 Dissertation New
92/97
85
Definition 6.11. (Cauchy sequence) A sequence {x n , n = 1, 2, } of elements of an
inner product space is said to be a Cauchy sequence if
||x n -x m || 0 as n, m .
For Definition 6.11, see Brockwell and Davis (1994, p. 45).
Definition 6.12. (Hilbert space) A Hilbert space H is an inner product space which is
complete, i.e. an inner product space in which every Cauchy sequence {x n }converges in
norm to some elementx H.
For Definitions 6.12 , see Brockwell and Davis (1994, p. 45).
Example 6.1. ( The standard inner product for C n ). The standard inner product for
C
n
is defined by
= =
n
i
ii yx1
_
,
wherex= (x1, x 2, , x n )andy= (y 1, y 2, , y n ).
Remark. C n is definedas the set of all n-tuples of complex numbers, that is
z =
nz
z
z
2
1
,
where z 1 , z 2 , , z n C.
8/11/2019 Dissertation New
93/97
86
For Examples 6.1, see Dorny (1980, p240).
Example 6.2. (The standard inner product for function spaces). The standard inner
product for a function space C(a, b) is defined by
= b
a
tgtf )()( dt,
where f and g C(a, b). Where C(a, b) is the set of continuous, complex-valued func-
tions on [a, b].
For Examples 6.2 , see Dorny (1980, p240-241).
Example 6.3 (A weighted Inner product for function spaces). An inner product of the
following form is often appropriate for C(a, b):
= b
a
tgtftw )()()( dt, 0 < w(t)< .
For Examples 6.3, see Dorny (1980, p242).
8/11/2019 Dissertation New
94/97
87
A6.3 Orthogonal Projection and Orthonormal Bases
Definition 6.13. ( Orthogonal ) Let V be a complex inner product space. Forx, yV,
we sayx andyare orthogonal, writtenx y, if = 0.
For Definition 6.13, see Frazier (1999, p84).
Definition 6.14. (Orthogonal and orthonormal set) Let V be a complex inner product
space and X be a set in V. X is called an orthogonal set if any two different vectors of X
are orthogonal; X is called an orthonormal set if X is an orthogonal set and ||x|| = 1 for
allx X.
For Definition 6.14, see Frazier (1999, p84).
Theorem 6.7. Let V be a complex inner product space, and let X = {x1, x 2, , x n } be
an orthogonal set of vectors in V and 0 X. Then X is a linearly independent set.
For Theorem 6.7 and its proof, see Frazier (1999, p84).
Theorem 6.8. Let V be a complex inner product space, and let X = {x 1, x 2, , x n } be
an orthogonal set in V withx j 0 for allj. Ifx span X, then
x= jn
j j
jx
x
xx
=
>