-
See discussions, stats, and author profiles for this publication
at: https://www.researchgate.net/publication/319750286
Nonlinear Methods for Design-Space Dimensionality Reduction in
Shape
Optimization
Chapter · September 2017
DOI: 10.1007/978-3-319-72926-8_11
CITATIONS
3READS
198
4 authors, including:
Some of the authors of this publication are also working on
these related projects:
FRIDA - FRamework for Integrated Design of Aircraft View
project
Multifidelity metamodels and adaptive grid refinement for shape
optimistion View project
Danny D'Agostino
Sapienza University of Rome
10 PUBLICATIONS 26
CITATIONS
SEE PROFILE
Andrea Serani
Italian National Research Council
59 PUBLICATIONS 348
CITATIONS
SEE PROFILE
Matteo Diez
Italian National Research Council
142 PUBLICATIONS 1,124
CITATIONS
SEE PROFILE
All content following this page was uploaded by Andrea Serani on
15 September 2017.
The user has requested enhancement of the downloaded file.
https://www.researchgate.net/publication/319750286_Nonlinear_Methods_for_Design-Space_Dimensionality_Reduction_in_Shape_Optimization?enrichId=rgreq-d205834d8db27954b5ac86c3feae6d86-XXX&enrichSource=Y292ZXJQYWdlOzMxOTc1MDI4NjtBUzo1Mzg3ODE5NTY1MzQyNzJAMTUwNTQ2NzAzNjkyOQ%3D%3D&el=1_x_2&_esc=publicationCoverPdfhttps://www.researchgate.net/publication/319750286_Nonlinear_Methods_for_Design-Space_Dimensionality_Reduction_in_Shape_Optimization?enrichId=rgreq-d205834d8db27954b5ac86c3feae6d86-XXX&enrichSource=Y292ZXJQYWdlOzMxOTc1MDI4NjtBUzo1Mzg3ODE5NTY1MzQyNzJAMTUwNTQ2NzAzNjkyOQ%3D%3D&el=1_x_3&_esc=publicationCoverPdfhttps://www.researchgate.net/project/FRIDA-FRamework-for-Integrated-Design-of-Aircraft?enrichId=rgreq-d205834d8db27954b5ac86c3feae6d86-XXX&enrichSource=Y292ZXJQYWdlOzMxOTc1MDI4NjtBUzo1Mzg3ODE5NTY1MzQyNzJAMTUwNTQ2NzAzNjkyOQ%3D%3D&el=1_x_9&_esc=publicationCoverPdfhttps://www.researchgate.net/project/Multifidelity-metamodels-and-adaptive-grid-refinement-for-shape-optimistion?enrichId=rgreq-d205834d8db27954b5ac86c3feae6d86-XXX&enrichSource=Y292ZXJQYWdlOzMxOTc1MDI4NjtBUzo1Mzg3ODE5NTY1MzQyNzJAMTUwNTQ2NzAzNjkyOQ%3D%3D&el=1_x_9&_esc=publicationCoverPdfhttps://www.researchgate.net/?enrichId=rgreq-d205834d8db27954b5ac86c3feae6d86-XXX&enrichSource=Y292ZXJQYWdlOzMxOTc1MDI4NjtBUzo1Mzg3ODE5NTY1MzQyNzJAMTUwNTQ2NzAzNjkyOQ%3D%3D&el=1_x_1&_esc=publicationCoverPdfhttps://www.researchgate.net/profile/Danny_Dagostino?enrichId=rgreq-d205834d8db27954b5ac86c3feae6d86-XXX&enrichSource=Y292ZXJQYWdlOzMxOTc1MDI4NjtBUzo1Mzg3ODE5NTY1MzQyNzJAMTUwNTQ2NzAzNjkyOQ%3D%3D&el=1_x_4&_esc=publicationCoverPdfhttps://www.researchgate.net/profile/Danny_Dagostino?enrichId=rgreq-d205834d8db27954b5ac86c3feae6d86-XXX&enrichSource=Y292ZXJQYWdlOzMxOTc1MDI4NjtBUzo1Mzg3ODE5NTY1MzQyNzJAMTUwNTQ2NzAzNjkyOQ%3D%3D&el=1_x_5&_esc=publicationCoverPdfhttps://www.researchgate.net/institution/Sapienza_University_of_Rome?enrichId=rgreq-d205834d8db27954b5ac86c3feae6d86-XXX&enrichSource=Y292ZXJQYWdlOzMxOTc1MDI4NjtBUzo1Mzg3ODE5NTY1MzQyNzJAMTUwNTQ2NzAzNjkyOQ%3D%3D&el=1_x_6&_esc=publicationCoverPdfhttps://www.researchgate.net/profile/Danny_Dagostino?enrichId=rgreq-d205834d8db27954b5ac86c3feae6d86-XXX&enrichSource=Y292ZXJQYWdlOzMxOTc1MDI4NjtBUzo1Mzg3ODE5NTY1MzQyNzJAMTUwNTQ2NzAzNjkyOQ%3D%3D&el=1_x_7&_esc=publicationCoverPdfhttps://www.researchgate.net/profile/Andrea_Serani?enrichId=rgreq-d205834d8db27954b5ac86c3feae6d86-XXX&enrichSource=Y292ZXJQYWdlOzMxOTc1MDI4NjtBUzo1Mzg3ODE5NTY1MzQyNzJAMTUwNTQ2NzAzNjkyOQ%3D%3D&el=1_x_4&_esc=publicationCoverPdfhttps://www.researchgate.net/profile/Andrea_Serani?enrichId=rgreq-d205834d8db27954b5ac86c3feae6d86-XXX&enrichSource=Y292ZXJQYWdlOzMxOTc1MDI4NjtBUzo1Mzg3ODE5NTY1MzQyNzJAMTUwNTQ2NzAzNjkyOQ%3D%3D&el=1_x_5&_esc=publicationCoverPdfhttps://www.researchgate.net/institution/Italian_National_Research_Council?enrichId=rgreq-d205834d8db27954b5ac86c3feae6d86-XXX&enrichSource=Y292ZXJQYWdlOzMxOTc1MDI4NjtBUzo1Mzg3ODE5NTY1MzQyNzJAMTUwNTQ2NzAzNjkyOQ%3D%3D&el=1_x_6&_esc=publicationCoverPdfhttps://www.researchgate.net/profile/Andrea_Serani?enrichId=rgreq-d205834d8db27954b5ac86c3feae6d86-XXX&enrichSource=Y292ZXJQYWdlOzMxOTc1MDI4NjtBUzo1Mzg3ODE5NTY1MzQyNzJAMTUwNTQ2NzAzNjkyOQ%3D%3D&el=1_x_7&_esc=publicationCoverPdfhttps://www.researchgate.net/profile/Matteo_Diez?enrichId=rgreq-d205834d8db27954b5ac86c3feae6d86-XXX&enrichSource=Y292ZXJQYWdlOzMxOTc1MDI4NjtBUzo1Mzg3ODE5NTY1MzQyNzJAMTUwNTQ2NzAzNjkyOQ%3D%3D&el=1_x_4&_esc=publicationCoverPdfhttps://www.researchgate.net/profile/Matteo_Diez?enrichId=rgreq-d205834d8db27954b5ac86c3feae6d86-XXX&enrichSource=Y292ZXJQYWdlOzMxOTc1MDI4NjtBUzo1Mzg3ODE5NTY1MzQyNzJAMTUwNTQ2NzAzNjkyOQ%3D%3D&el=1_x_5&_esc=publicationCoverPdfhttps://www.researchgate.net/institution/Italian_National_Research_Council?enrichId=rgreq-d205834d8db27954b5ac86c3feae6d86-XXX&enrichSource=Y292ZXJQYWdlOzMxOTc1MDI4NjtBUzo1Mzg3ODE5NTY1MzQyNzJAMTUwNTQ2NzAzNjkyOQ%3D%3D&el=1_x_6&_esc=publicationCoverPdfhttps://www.researchgate.net/profile/Matteo_Diez?enrichId=rgreq-d205834d8db27954b5ac86c3feae6d86-XXX&enrichSource=Y292ZXJQYWdlOzMxOTc1MDI4NjtBUzo1Mzg3ODE5NTY1MzQyNzJAMTUwNTQ2NzAzNjkyOQ%3D%3D&el=1_x_7&_esc=publicationCoverPdfhttps://www.researchgate.net/profile/Andrea_Serani?enrichId=rgreq-d205834d8db27954b5ac86c3feae6d86-XXX&enrichSource=Y292ZXJQYWdlOzMxOTc1MDI4NjtBUzo1Mzg3ODE5NTY1MzQyNzJAMTUwNTQ2NzAzNjkyOQ%3D%3D&el=1_x_10&_esc=publicationCoverPdf
-
Nonlinear Methods for Design-SpaceDimensionality Reduction in
Shape Optimization
(Pre-print)
Danny D’Agostino1,2, Andrea Serani1, Emilio F. Campana1, and
Matteo Diez1
1 CNR-INSEAN, Natl. Research Council–Marine Tech. Research
Inst., Rome, [email protected]
2 Department of Computer, Control, and Management Engineering
“A. Ruberti”,Sapienza University of Rome, Rome, Italy
Abstract. In shape optimization, design improvements
significantly de-pend on the dimension and variability of the
design space. High dimen-sional and variability spaces are more
difficult to explore, but also usuallyallow for more significant
improvements. The assessment and breakdownof design-space
dimensionality and variability are therefore key elementsto shape
optimization. A linear method based on the principal compo-nent
analysis (PCA) has been developed in earlier research to build
areduced-dimensionality design-space, resolving the 95% of the
originalgeometric variance. The present work introduces an
extension to moreefficient nonlinear approaches. Specifically the
use of Kernel PCA, Lo-cal PCA, and Deep Autoencoder (DAE) is
discussed. The methods aredemonstrated for the design-space
dimensionality reduction of the hullform of a USS Arleigh
Burke-class destroyer. Nonlinear methods areshown to be more
effective than linear PCA. DAE shows the best per-formance
overall.
Keywords: Shape optimization, hull-form design, nonlinear
dimension-ality reduction, kernel methods, deep autoencoder
1 Introduction
The simulation-based design (SBD) paradigm has demonstrated its
capabilityof supporting the design decision process, providing
large sets of design optionsand reducing time and costs of the
design process. The recent development ofhigh performance computing
(HPC) systems has driven the SBD towards itsintegration with
optimization algorithms, moving the SBD paradigm further,to
automatic SBD optimization (SBDO). In shape optimization, SBDO
consistsof three main elements: (i) a simulation tool, (ii) an
optimization algorithm,and (iii) a shape modification tool, which
need to be integrated efficiently androbustly. In this context,
design improvements significantly depend on the dimen-sion and
extension of the design space: high dimensional and variability
spacesare more difficult and computationally expensive to explore
but, at the sametime, potentially allow for bigger improvements.
The assessment and breakdown
-
2
of the design-space dimensionality and variability are therefore
a key elementfor the success of the SBDO [1].
Online linear dimensionality reduction techniques have been
developed, re-quiring the evaluation of the objective function or
its gradient. As an example,principal component analysis (PCA) or
proper orthogonal decomposition (POD)methods have been applied for
reduced-dimensionality local representations offeasible design
regions [2]. A PCA/POD based approach is used in the active
sub-space method (ASM) [3] to discover and exploit low-dimensional
and monotonictrends in the objective function, based on the
evaluation of its gradient. Onlinemethods improve the shape
optimization efficiency by basis rotation and/or di-mensionality
reduction. Nevertheless, they do not provide an assessment of
thedesign space and the associated shape parametrization before
optimization isperformed or objective function and/or gradient are
evaluated.
Offline linear methodologies have been developed with focus on
design-spacevariability and dimensionality reduction for efficient
optimization procedures. Amethod based on the Karhunen-Loève
expansion (KLE) has been formulatedfor the assessment of the shape
modification variability and the definition ofa
reduced-dimensionality global model of the shape modification
vector in [1].No objective function evaluation nor gradient is
required by the method. TheKLE is applied to the continuous shape
modification vector, requiring the so-lution of a Fredholm integral
equation of the second kind. Once the equationis discretized, the
problem reduces to the PCA of discrete data. Offline linearmethods
improve the shape optimization efficiency by reparametrization
anddimensionality reduction, providing the assessment of the design
space and theshape parametrization before optimization and/or
performance analysis are per-formed. The assessment is based on the
geometric variability associated to thedesign space of the shape
optimization. Although linear methods have been suc-cessfully
applied for a wide range of problems, they may be not efficient
whencomplex non linear relationship are involved in the performance
analysis andoptimization.
In the last years researchers have developed nonlinear methods
for data di-mensionality reduction. Nonlinear dimensionality
reduction (NLDR) methodsgeneralize linear methods to address data
with nonlinear structures. Kernel PCA(KPCA) solves a PCA
eigenproblem in a new space (called feature space) byusing kernel
methods [4]. Local PCA (LPCA) divides the initial design spacein k
clusters and a PCA is applied for each of them, supposing that the
datain each cluster has an approximate linear structure. LPCA
techniques may bedifferentiated based on the clustering method,
which may follow k-means [5] orspectral approaches [6]. Artificial
neural networks (ANN) have been also used toreduce data
dimensionality, by performing both encoder and decoder tasks
(themethod is also known as autoencoder).
The objective of the present work is to combine NLDR techniques
with shapeparametrization in SBDO for ship hydrodynamics.
Specifically KPCA, LPCAwith k-means (LPCA-KM), LPCA with spectral
clustering (LPCA-SC), andDeep Autoencoder (DAE) are used to build a
reduced-dimensionality design-
-
3
space, resolving at least the 95% of the original design
variability based onthe concept of geometric variance [1]. The
methods are demonstrated for thedesign-space dimensionality
reduction of the hull form of USS Arleigh Burke-class destroyer,
namely the DTMB 5415 model, an early and open to publicversion of
the DDG-51. The effectiveness of the NLDR techniques is shown
anddiscussed, comparing the results to the linear KLE/PCA method
from earlierwork [1].
2 Dimensionality Reduction Methods
General definitions and assumptions for the current problem are
presented in thefollowing, along with linear and nonlinear
dimensionality reduction methods.
2.1 General Definitions and Assumptions
Consider a geometric domain G (which identifies the initial
shape) and a set ofcoordinates x ∈ G.
!
Fig. 1: Scheme and notation for thecurrent formulation, showing
an ex-ample for n = 1 and m = 2
Assume that u ∈ U is the design vari-able vector, which defines
a continuousshape modification vector δ(x,u). Con-sider the design
variables u as a randomfield defined over a domain U , with
asso-ciated probability density function p(u).The associated mean
shape modificationis evaluated as
〈δ〉 =∫Uδ(x,u)p(u)du (1)
If one defines the internal product inG as
(f ,g) =
∫G
f(x) · g(x) dx (2)
with associated norm ‖f‖ = (f , f)1/2, the variance associated
to the shape mod-ification vector (geometric variance) may be
defined as
σ2 =〈‖δ̂‖2
〉=
∫U
∫Gδ̂(x,u) · δ̂(x,u)p(u)dxdu (3)
where δ̂ = δ−〈δ〉, and 〈·〉 denotes the ensemble average over u.
Generally, x ∈ Rnwith n = 1, 2, 3, u ∈ RM with M number of design
variables, and δ ∈ Rm withm = 1, 2, 3 (with m not necessarily equal
to n). Figure 1 shows an examplewith n = 1 and m = 2. Ensemble
averages 〈·〉 over u ∈ U may be evaluated by
-
4
Monte Carlo (MC) sampling using a statistically convergent
number of randomrealizations S, {uk}Sk=1 ∼ p(u). These are
collected in a [S × L] matrix
D =
d(u1) . . . d(uS)T (4)
representing the (MC sampled) original design space, where d(uk)
={dq(uk)}mq=1 is the deviation from the mean of the shape
modification vectorand its q-th component is evaluated at discrete
coordinates xt, t = 1 . . . , T , as
dq(uk) =
δq(x1,uk)
...δq(xT ,uk)
− 1SS∑k=1
δq(x1,uk)
...δq(xT ,uk)
(5)with δq = δ · eq, where {eq}mq=1 ∈ Rm is a basis of
orthogonal unit vector. Notethat L = mT .
A reduced-dimensionality representation of D is sought after for
later use inthe SBDO.
2.2 Principal Component Analysis
PCA allows to reduce the input dimensionality of the data,
performing a pro-jection of the points in a new linear subspace,
defined by the eigenvectors of the[L× L] covariance matrix C =
DTD/S. These eigenvectors have the propertiesto maximize the
variance of points projected on them and to minimize the
meansquared distance between the original points and the relative
projections [7]. Theprincipal components are defined by the
solution of the eigenproblem
Cz = λz (6)
The solutions {zi}Li=1 of the Eq. 6 are used to build a
reduced-dimensionalityspace for the shape modification vector d
as
d ≈N∑i=1
αizi = d̂ (7)
where αi is the i-th component of the new design variable vector
α ∈ RN .Equation 7 may be truncated to the N -th order, preserving
a desired level ofconfidence β (0 < β ≤ 1), provided that
N∑i=1
λi ≥ βL∑i=1
λi = βσ2 (8)
assuming λi ≥ λi+1. Only M eigenvalues are expected to be non
zeros.
-
5
2.3 Kernel Principal Component Analysis
The kernel PCA (KPCA) method [4] is a nonlinear extension of
PCA. It findsdirections of maximum variance in a higher (possibly
infinite) dimensional fea-ture space F , mapping the points from
the input space I by a possible nonlinearfunction Φ : I → F as
dk → Φ(dk), ∀k = 1, . . . , S (9)
where, for the sake of simplicity, the d(uk) of Eq. 4 is here
simplified in dk.Then PCA is computed in the feature space F .
Assuming that ∑k Φ(dk) = 0,the kernel principal component {zp}Pp=1
can be find solving the eigenproblem
ΣΦzp = λpzp (10)
where ΣΦ is the [P × P ] covariance matrix in the feature space
F , defined as
ΣΦ =1
S
S∑k=1
Φ(dk)Φ(dk)T (11)
KPCA allows the solution of Eq. 10 without computing explicitly
the Eq.9, since it appears only within an inner product [8], which
can be computedefficiently by a kernel function K(di,dk) =
Φ(di)
TΦ(dk). Defining zp as a linearexpansion of Φ(dk)
zp =
S∑k=1
cpkΦ(dk) (12)
the Eq. 10 can be recasted as
Kcp = λpScp (13)
where K is the symmetric and positive-semidefinite [S × S]
kernel matrix, withKik = K(di,dk). The length of the S-component
vector cp is chosen such thatzTp zp = λpSc
Tp cp = 1. Once the eigenproblem in Eq. 13 is solved, the new
design
variables can be found projecting Φ(d) on zp as
α = Φ(d)zp =
S∑k=1
cpkΦ(d)TΦ(dk) =
S∑k=1
cpkK(d,dk) (14)
The reconstruction of the original data from the feature space F
in KPCAis more problematic than PCA, since it needs to find, for
every point Φ(dk), therelative pre-image dk in the input space I.
In this paper, approximate pre-imagestechnique proposed in [9] is
used.
-
6
2.4 Local Principal Component Analysis
Local PCA (LPCA) performs a PCA for every different disjoint
region of theinput space I, assuming that, if the local regions are
small enough, the datamanifold will not curve much over the extent
of the region and the linear modelwill be a good fit [5].
The first step in LPCA is to cluster the data in k sets,
applying a clusteringalgorithm, such that D = {D1, . . . ,Di}ki=1.
Herein, LPCA is performed with twoclustering techniques: the
k-means (LPCA-KM) algorithm [10] and a spectralclustering (LPCA-SC)
[11]. The k-means clustering algorithm is described inAlg. 1.
Algorithm 1 k-means clustering algorithm
Require: Random k centroids as representative points of each
cluster Di ∀i = 1, . . . , k.1: repeat2: Assign each point dj to
the nearest centroid µi using the Euclidean distance as
similarity
measure.3: Update the centroids according to: µi =
1|Di|
∑dj∈Di
dj
4: until µi ∀i = 1, . . . , k remains unchanged
One issue in k-means is that using the euclidean distance as
similarity mea-sure assumes a convex shape to the underlying
clusters [12].
Spectral clustering can be effective even if the clusters shape
are more com-plex. There are several versions of the spectral
clustering algorithms, the maindifference is in which graph
Laplacian is used [6]. Herein, the symmetric nor-
malized Laplacian Asym = I − B−12 WB−
12 [11] is used and the corresponding
algorithm is summarized in Alg. 2 [6].After the data are
partitioned in k clusters, a PCA is performed on them
solving k PCA eigenproblem
Cizi = λizi ∀i = 1, . . . , k (15)LPCA results are highly
dependent by the clustering procedure and specially
by the number of clusters used. Moreover, the number of clusters
k should beset carefully to avoid extensive computation.
Algorithm 2 Normalized Spectral Clustering
Require: Let k the number of clusters to identify, build a
similarity graph as:
– K-nearest neighbor graphs: fix K, di is connected to a point
dj if it is among the K-nearestneighbor of di or viceversa.
1: Compute the adjacency matrix W of the graph and the diagonal
degree matrix B, where eachelement is equal to bii =
∑Sj=1 wij .
2: Compute the symmetric normalized Laplacian Asym.3: Find the
first k eigenvector v1, . . . ,vk corresponding to the k smallest
eigenvalues of Asym.4: Construct a [S × k] matrix V with the
eigenvectors as columns.5: Normalize the rows of matrix V by v̂ij =
vij/(
∑k v
2ik)
12
6: Run k-means on matrix V.
-
7
2.5 Deep Autoencoders
An autoencoder (AE) is an ANN that performs two main tasks [13]:
(1) anencoder function E maps the data d to compress data α; (2) a
decoder function Dmaps from the compressed data α back to d̂. This
operation is performed settingthe same number of neurons L in the
input and output layer and constrainingthe hidden layer to have N
< M neurons.
Consider a single hidden layer AE, if the new design variable α
can be writtenas
α = E(H(1)d + b(1)) (16)where H is a relative weight matrix, b
the bias vector, and the apex “(1)”
represent the hidden layer, then the reconstruction vector d̂
from α can beexpressed as
d̂ = D(H(2)α+ b(2)) (17)where the apex “(2)” represent the
output layer. The network parameters H andb, are evaluated
minimizing the reconstruction error
E(H(1),b(1),H(2),b(2)) =1
2
S∑k=1
||dk − d̂k||2 (18)
=1
2
S∑k=1
||dk −D(H(2)E(H(1)dk + b(1)) + b(2))||2
Fig. 2: Example of AE with onehidden layer with L = 3 andN =
2
If E and D are linear then the Eq. 18 has aunique global
minimum, in which the weightsin the hidden layer span the same
subspace asthe first N -principal components of the data[14, 15].
AE with nonlinear activation func-tions and more hidden layers
(called deep au-toencoder, DAE) provides a nonlinear
gener-alization of the PCA [16], but in this case theerror function
(Eq. 18) becomes non convexand the optimization algorithm may get
stuckin poor local minima. Moreover, the intrinsicdimensionality of
the data (the number of neurons N in the hidden layer) cannotbe
known a priori and have to be fixed respect to the reconstruction
error.
3 Shape Modification of a Destroyer Hull
The DTMB 5415 model is an open-to-public early concept of the
DDG-51, a USSArleigh Burke-class destroyer, widely used for both
towing tank experiments [17]and hull-form SBDO [18]. Figure 3 shows
its geometry and body surface gridused to discretize the shape
modification domain.
-
8
The offline design-space assessment and dimensionality reduction
of the DTMB5415 hull form (assuming full-scale with a length
between perpendiculars Lpp =142 m) is presented as a
pre-optimization study of the following problem
Minimize f(u)subject to ga(u) = 0, with a = 1, . . . , A
and to he(u) ≤ 0, with e = 1, . . . , E(19)
where f is the objective function related to the ship
performance (i.e. resistance,seakeeping, etc.) and u are the
(original) design variables. Geometrical equalityconstraints, ga,
include fixed length between perpendicular (Lpp) and displace-ment
(∇), whereas geometrical inequality constraints, he, include 5%
maximumvariation of beam and draught and reserved volume for the
sonar in the bowdome, corresponding to 4.9 m diameter and 1.7 m
length (cylinder).
X Y
ZI
J
Fig. 3: DTMB 5415 geometry and bodysurface discretization
Shape modifications δ(x,u) areapplied directly on the
Cartesiancoordinates g of the computationalbody surface grid, as
per
g(u) = g0 + δ(x,u) (20)
where g0 represents the original grid.The shape modification is
defined
using a linear combination of M = 27vector-valued functions of
the Carte-sian coordinates x over a hyper-rectangle embedding the
demi hull [18]
ψi(x) : V = [0, Lx1 ]× [0, Lx2 ]× [0, Lx3 ] ∈ R3 −→ R3 (21)with
i = 1, ...,M , as
δ(x,u) =
M∑i=1
uiψi(x) (22)
where the coefficients ui ∈ R (i = 1, . . . ,M) are the
(original) design variables,
ψi(x) :=
3∏j=1
sin
(aijπxjLxj
+ rij
)eq(i) (23)
and the following orthogonality property is imposed:∫Vψi(x)
·ψk(x)dx = δik (24)
In Eq. 23, {aij}3j=1 ∈ R define the order of the function along
j-th axis;{rij}3j=1 ∈ R are the corresponding spatial phases;
{Lxj}3j=1 are the hyper-rectangle edge lengths; eq(i) is a unit
vector. Modifications are applied along x1,x2, or x3, with q(i) =
1, 2, or 3 respectively. The parameter values used here aretaken
from [18].
Fixed Lpp and ∇ are satisfied by automatic geometric scaling,
while geome-tries exceeding the constraints are not considered.
-
9
4 Numerical Results
The results obtained by linear PCA and the nonlinear methods
(KPCA, LPCA-KM, LPCA-SC, and DAE) are presented in the following
subsections. Two eval-uation metrics are used to assess the
methods’ performance and compare them.
4.1 Evaluation Metrics
The methods are assessed by the portion of original geometric
variance resolved(β̂) and the root mean square error (RMSE) of
matrix reconstruction D̂, definedas
β̂ =1S
∑Lj=1
∑Sk=1(d̂jk − µ̂j)2
1S
∑Lj=1
∑Sk=1(djk − µj)2
and RMSE =
√√√√ 1S
S∑k=1
||dk − d̂k||2 (25)
where µ̂j is the mean value of D̂ j-th column.
4.2 Evaluation of Design-Space Dimensionality
ReductionCapabilities
In assessing the methods’ performance, a cubic polynomial kernel
is used for theKPCA, a number of cluster k = 32 and 24 is used for
LPCA-KM and LPCA-SCrespectively, a seven hidden layer DAE (composed
by 300-150-50-N -50-150-300neurons) with hyperbolic tangent (as
activation function) is used and trainedwith Adam optimization
algorithm [19].
Table 1: Numerical results
Method N [–] β̂% RMSE/Lpp
PCA 24 95.0 1.12E-1KPCA 18 100. 0.00E+0
LPCA-KM 12 95.0 1.12E-1LPCA-SC 15 95.4 1.08E-1
DAE 5 97.8 9.60E-2
The design space (M = 27) is sam-pled using a uniform random
distributionof S = 1, 000 hull-form designs. For
eachdimensionality-reduction method, Fig. 4ashows the geometric
variance (β̂%) resolvedby a N -dimensional design space,
whereasFig. 4b shows the corresponding reconstruc-tion error
(RMSE). The nonlinear methods
result to be more effective than the linear PCA in terms of both
β̂% and RMSE.Specifically, in order to reduce the design-space
dimensionality while resolving
at least the 95% of the original geometric variance, N = 24 is
required by PCA,whereas N = 18, 12, 15, and 5 are needed by KPCA,
LPCA-KM, LPCA-SC, andDAE, respectively. The results are summarized
in Tab. 1. It is worth noting thatKPCA requires N = 18, but
resolves the 100% of the original variance and showsa
reconstruction error equal to zero. In the current study, it was
not possibleto reduce N further, due to numerical issue associated
to the computation ofpre-images.
Finally, Fig. 5 shows the shape modification (δy) and the
reconstruction error(∆δy) versus grid-node index (I, J), and the
corresponding hull stations for adesign originally included in the
data matrix D. For this design, LPCA shows
-
10
the largest reconstruction error. PCA and DAE produce a close
reconstructionto the target, whereas KPCA reproduce the target
exactly. With only N = 5,DAE is the most efficient overall.
5 Conclusions and Future Work
Four nonlinear methods for design-space dimensionality-reduction
in shape opti-mization have been presented and compared.
Specifically, kernel PCA (KPCA),local PCA with k-means and spectral
clustering (respectively LPCA-KM andLPCA-SC), and deep autoencoder
(DAE) have been used for an offline pre-optimization
dimensionality-reduction of the hull-form parametrization of
theDTMB 5415 model hull. A linear PCA method from earlier studies
has beenalso included in the analysis, for comparison.
The original shape parametrization was defined by M = 27 design
variables.The reduced-dimensionality space is required to resolve
at least the 95% of theoriginal design variability, based on the
concept of geometric variance. The linearPCA achieved a reduction
of 11.2% of the original design dimensionality (requir-ing a number
of design variables N = 24). All nonlinear methods outperform
thelinear PCA. Specifically, a 33.4% dimensionality reduction is
achieved by KPCA(N = 18), 55.5% by LPCA-KM (N = 12), 44.4% by
LPCA-SC (N = 15), andfinally a remarkable 81.5% by DAE (N = 5).
Nonlinear methods have showntheir superior effectiveness in terms
of both variance resolved and reconstructionerror, compared to
linear PCA. DAE have shown the best performance overall.
The analysis of some specific behavior of the methods presented,
such asthe assessment of the clusters used by the LPCA, will be
addressed in futurework. Moreover, in order to investigate further
on the methods’ effectiveness,future work will include the
optimization of the DTMB 5415 using the reduced-dimensionality
space produced by linear and nonlinear methods, with compar-ison of
objective function improvement and convergence to the optimum.
Also,combined geometry and physics based design variability studies
[20, 21] will beaddressed using current nonlinear methods.
0 5 10 15 20 25N [−]
0.0
0.2
0.4
0.6
0.8
1.0
β[-
]
PCA
KPCA
LPCA-KM
LPCA-SC
DAE
β = 0.95 [-]
(a) Geometric variance resolved
0 5 10 15 20 25N [−]
0.0
0.1
0.2
0.3
0.4
0.5
RMSE
[m]
PCA
KPCA
LPCA-KM
LPCA-SC
DAE
(b) Reconstruction RMSE
Fig. 4: Convergence of dimensionality-reduction methods in terms
of β̂% (a)and RMSE (b) versus the reduced-dimensionality N
-
11
Geometry modification Reconstruction error Hull stations
Target(M = 27)
0 10 20 30 40 50 60 70 80I[-]
0
5
10
15
20
J[-]
−3−2−1
0
1
2
3
δ y[m
]
PCA(N = 24)
0 10 20 30 40 50 60 70 80I[-]
0
5
10
15
20
J[-]
−3−2−1
0
1
2
3
δ y[m
]
0 10 20 30 40 50 60 70 80I[-]
0
5
10
15
20
J[-]
−1.2−0.8−0.4
0.0
0.4
0.8
1.2
∆δ y
[m]
Y [m]
-10 -5 0 5 10
Z [m
]
-5
0
5
10TargetReconst.
WL
KPCA(N = 18)
0 10 20 30 40 50 60 70 80I[-]
0
5
10
15
20
J[-]
−3−2−1
0
1
2
3
δ y[m
]
0 10 20 30 40 50 60 70 80I[-]
0
5
10
15
20
J[-]
−1.2−0.8−0.4
0.0
0.4
0.8
1.2
∆δ y
[m]
Y [m]
-10 -5 0 5 10
Z [m
]
-5
0
5
10TargetReconst.
WL
LPCA-KM(N = 12)
0 10 20 30 40 50 60 70 80I[-]
0
5
10
15
20
J[-]
−3−2−1
0
1
2
3
δ y[m
]
0 10 20 30 40 50 60 70 80I[-]
0
5
10
15
20
J[-]
−1.2−0.8−0.4
0.0
0.4
0.8
1.2
∆δ y
[m]
Y [m]
-10 -5 0 5 10
Z [m
]
-5
0
5
10TargetReconst.
WL
LPCA-SC(N = 15)
0 10 20 30 40 50 60 70 80I[-]
0
5
10
15
20
J[-]
−3−2−1
0
1
2
3
δ y[m
]
0 10 20 30 40 50 60 70 80I[-]
0
5
10
15
20
J[-]
−1.2−0.8−0.4
0.0
0.4
0.8
1.2∆δ y
[m]
Y [m]
-10 -5 0 5 10
Z [m
]
-5
0
5
10TargetReconst.
WL
DAE(N = 5)
0 10 20 30 40 50 60 70 80I[-]
0
5
10
15
20
J[-]
−3−2−1
0
1
2
3
δ y[m
]
0 10 20 30 40 50 60 70 80I[-]
0
5
10
15
20
J[-]
−1.2−0.8−0.4
0.0
0.4
0.8
1.2
∆δ y
[m]
Y [m]
-10 -5 0 5 10
Z [m
]
-5
0
5
10TargetReconst.
WL
Fig. 5: Reconstruction of the geometry modification vector δy,
reconstructionerror, and corresponding hull stations of target
geometry (original input)
Acknowledgments. The work is supported by the US Office of Naval
ResearchGlobal, NICOP grant N62909-15-1-2016, under the
administration of Dr Woei-Min Lin, Dr. Salahuddin Ahmed, and Dr.
Ki-Han Kim, and by the Italian Flag-ship Project RITMARE. The
research is performed within NATO STO TaskGroup AVT-252 Stochastic
Design Optimization for Naval and Aero MilitaryVehicles. The
authors wish to thank Prof. Frederick Stern and Dr. Manivan-nan
Kandasamy of The University of Iowa for inspiring the current
research on
-
12
nonlinear dimensionality reduction methods.
References
1. Diez, M., Campana, E.F., Stern, F.: Design-space
dimensionality reduction inshape optimization by Karhunen–Loève
expansion. Computer Methods in AppliedMechanics and Engineering 283
(2015) 1525–1544
2. Raghavan, B., Breitkopf, P., Tourbier, Y., Villon, P.:
Towards a space reduction ap-proach for efficient structural shape
optimization. Structural and MultidisciplinaryOptimization 48
(2013) 9871000
3. Lukaczyk, T., Palacios, F., Alonso, J.J., Constantine, P.:
Active subspaces forshape optimization. In: Proceedings of the 10th
AIAA Multidisciplinary DesignOptimization Specialist Conference,
National Harbor, Maryland, USA, 13-17 Jan-uary. (2014)
4. Schölkopf, B., Smola, A., Müller, K.R.: Nonlinear component
analysis as a kerneleigenvalue problem. Neural computation 10(5)
(1998) 1299–1319
5. Kambhatla, N., Leen, T.K.: Dimension reduction by local
principal componentanalysis. Neural computation 9(7) (1997)
1493–1516
6. Von Luxburg, U.: A tutorial on spectral clustering.
Statistics and computing 17(4)(2007) 395–416
7. Bishop, C.M.: Pattern Recognition and Machine Learning
(Information Scienceand Statistics). Springer-Verlag New York,
Inc., Secaucus, NJ, USA (2006)
8. Smola, A.J., Schölkopf, B.: Learning with kernels. Citeseer
(1998)
9. Bakır, G.H., Weston, J., Schölkopf, B.: Learning to find
pre-images. Advances inneural information processing systems 16
(2004) 449–456
10. Lloyd, S.: Least squares quantization in PCM. IEEE
transactions on informationtheory 28(2) (1982) 129–137
11. Ng, A.Y., Jordan, M.I., Weiss, Y., et al.: On spectral
clustering: Analysis and analgorithm. In: NIPS. Volume 14. (2001)
849–856
12. Aggarwal, C.C., Reddy, C.K.: Data clustering: algorithms and
applications. Chap-man and Hall/CRC (2013)
13. Hinton, G.E., Salakhutdinov, R.R.: Reducing the
dimensionality of data withneural networks. Science 313(5786) (July
2006) 504–507
14. Bourlard, H., Kamp, Y.: Auto-association by multilayer
perceptrons and singularvalue decomposition. Biological cybernetics
59(4) (1988) 291–294
15. Baldi, P., Hornik, K.: Neural networks and principal
component analysis: Learningfrom examples without local minima.
Neural networks 2(1) (1989) 53–58
16. LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. Nature
521(7553) (2015) 436–444
17. Stern, F., Longo, J., Penna, R., Olivieri, A., Ratcliffe,
T., Coleman, H.: Inter-national collaboration on benchmark CFD
validation data for surface combatantDTMB model 5415. In:
Proceedings of the Twenty-Third Symposium on NavalHydrodynamics,
Val de Reuil, France, September 17-22 (2000)
18. Serani, A., Fasano, G., Liuzzi, G., Lucidi, S., Iemma, U.,
Campana, E.F., Stern, F.,Diez, M.: Ship hydrodynamic optimization
by local hybridization of deterministicderivative-free global
algorithms. Applied Ocean Research 59 (2016) 115 – 128
19. Kingma, D., Ba, J.: Adam: A method for stochastic
optimization. arXiv preprintarXiv:1412.6980 (2014)
-
13
20. Diez, M., Serani, A., Stern, F., Campana, E.F.: Combined
geometry and physicsbased method for design-space dimensionality
reduction in hydrodynamic shapeoptimization. In: Proceedings of the
31st Symposium on Naval Hydrodynamics,Monterey, CA, USA. (2016)
21. Serani, A., Campana, E.F., Diez, M., Stern, F.: Towards
augmented design-spaceexploration via combined geometry and physics
based Karhunen-Loève expansion.In: 18th AIAA/ISSMO
Multidisciplinary Analysis and Optimization Conference(MA&O),
AVIATION 2017, Denver, USA, June 5-9 (2017)
View publication statsView publication stats
https://www.researchgate.net/publication/319750286