Top Banner
Dongwook Lee Applied Mathematics & Statistics University of California, Santa Cruz High-order Numerical Methods for Predictive Science on Large-scale High-performance Computing Architectures Mathematics Colloquium, November 18, 2014 FLASH Simulation of a 3D Core-collapse Supernova Courtesy of S. Couch MIRA, BG/Q, Argonne National Lab 49,152 nodes, 786,432 cores OMEGA Laser (US)
61

Mathematics Colloquium, UCSC

Jul 24, 2015

Download

Education

dongwook159
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Mathematics Colloquium, UCSC

Dongwook LeeApplied Mathematics & Statistics

University of California, Santa Cruz

High-order Numerical Methods for Predictive Science on

Large-scale High-performanceComputing Architectures

Mathematics Colloquium, November 18, 2014

FLASH Simulation of a 3D Core-collapse SupernovaCourtesy of S. Couch

MIRA, BG/Q, Argonne National Lab49,152 nodes, 786,432 cores OMEGA Laser (US)

Page 2: Mathematics Colloquium, UCSC

Theory

Experiment ScientificComputation

validation

verification

Cyclic Relationship

Page 3: Mathematics Colloquium, UCSC

Example: Supersonic Airflow

Theory

Experiment Simulation

Page 4: Mathematics Colloquium, UCSC

Topics for Today

1. High Performance Computing

2. Ideas on Numerical Methods

3. Validation & Predictive Science

Page 5: Mathematics Colloquium, UCSC

First Episode

1. High Performance Computing

2. Ideas on Numerical Methods

3. Validation & Predictive Science

Page 6: Mathematics Colloquium, UCSC

High Performance Computing (HPC)

‣To solve large problems in science, engineering, or

business

‣Modern HPC architectures have

▪ increasing number of cores

▪ declining memory/core

‣This trend will continue for the foreseeable future

Page 7: Mathematics Colloquium, UCSC

High Performance Computing (HPC)

‣This tension between computation & memory brings a

paradigm shift in numerical algorithms for HPC

‣To enable scientific computing on HPC architectures:

▪ efficient parallel computing, (e.g., data parallelism, task

parallelism, MPI, multi-threading, GPU accelerator, etc.)

▪ better numerical algorithms for HPC

Page 8: Mathematics Colloquium, UCSC

Numerical Algorithms for HPC

‣Numerical algorithms should conform to the

abundance of computing power and the scarcity of

memory.

‣But…

▪ without losing solution accuracy.

▪ with maintaining maximum solution stability.

▪ with faster convergence to a “correct” solution.

Page 9: Mathematics Colloquium, UCSC

High-Order Numerical Algorithms

‣ A good solution to this is to use high-order algorithms.

‣ They provide more accurate numerical solutions using

▪ less grid points (=memory save).

▪ higher-order mathematical approximations (promoting

more floating point operations, or computation).

▪ faster convergence to solution.

Page 10: Mathematics Colloquium, UCSC

Large Scale Astrophysics Codes

▪ FLASH (Flash group, U of Chicago)

▪ PLUTO (Mignone, U of Torino),

▪ CHOMBO (Colella, LBL)

▪ CASTRO (Almgren, Woosley, LBL, UCSC)

▪ MAESTRO (Zingale, Almgren, SUNY, LBL)

▪ ENZO (Bryan, Norman, Abel, Enzo group)

▪ BATS-R-US (CSEM, U of Michigan)

▪ RAMSES (Teyssier, CEA)

▪ CHARM (Miniati, ETH)

▪ AMRVAC (Toth, Keppens, CPA, K.U.Leuven)

▪ ATHENA (Stone, Princeton)

▪ ORION (Klein, McKee, U of Berkeley)

▪ ASTROBear (Frank, U of Rochester)

▪ ART (Kravtsov, Klypin, U of Chicago)

▪ NIRVANA (Ziegler, Leibniz-Institut für Astrophysik Potsdam), and others

Peta- ScaleCurrent HPC

Future HPC

Giga- ScaleCurrent Laptop/Desktop

?

Page 11: Mathematics Colloquium, UCSC

The FLASH Code

‣FLASH is is free, open source code for astrophysics and HEDP.

▪ modular, multi-physics, adaptive mesh refinement (AMR), parallel (MPI & OpenMP), finite-volume Eulerian compressible code for solving hydrodynamics and MHD

▪ professionally software engineered and maintained (daily regression test suite, code verification/validation), inline/online documentation

▪ 8500 downloads, 1500 authors, 1000 papers

▪ FLASH can run on various platforms from laptops to supercomputing (peta-scale) systems such as IBM BG/P and BG/Q.

Page 12: Mathematics Colloquium, UCSC

Scientific Simulations using FLASH

cosmological cluster formation supersonic MHD

turbulence

Type Ia SN

RT

CCSN

ram pressure stripping

laser slab

rigid body structure

Accretion Torus

LULI/Vulcan experiments: B-field generation/amplification

Page 13: Mathematics Colloquium, UCSC

Parallel Computing

‣ Adaptive Mesh Refinement (w/ Paramesh)

▪ conventional parallelism via MPI (Message

Passing Interface)

▪ domain decomposition distributed over

multiple processor units

▪ distributed memory (cf. shared memory)

uniform grid octree-basedblock AMR

patch-based AMR

Single block

Page 14: Mathematics Colloquium, UCSC

Parallelization, Optimization & Speedup

16 cores/node

16 GB/node

4 threads/core

Page 15: Mathematics Colloquium, UCSC

▪ 5 leaf blocks in a single MPI rank▪ 2 threads/core (or 2 threads/rank)

thread block list thread within block

‣Multi-threading (shared memory) using OpenMP directives

▪more parallel computing on BG/Q using hardware threads on a core

▪ 16 cores/node, 4 threads/core

Parallelization, Optimization & Speedup

Page 16: Mathematics Colloquium, UCSC

FLASH Scaling Test on BG/Q

VESTA:2 racks

1024 nodes/rack

MIRA:48 racks

1024 nodes/rack

BG/Q:16 cores/node4 threads/core

16GB/core

512(4k)

1024(8k)

2048(16k)

4096(32k)

8192(64k)

16384(128k)

32768(256k)

Mira nodes (ranks), 8 threads/rank

0

2

4

6

8

10

10�

8co

reho

urs/

zone

/ste

p

0

5

10

15

20

25

10�

7co

reho

urs/

zone

/IO

wri

te

Total evolutionMHDGravityGridIO (1 write)⌫-LeakageIdeal

RTFlame, strong scaling: 4 threads/rank CCSN, weak scaling: 8 threads/rank

Page 17: Mathematics Colloquium, UCSC

Second Episode

1. High Performance Computing

2. Ideas on Numerical Methods

3. Validation & Predictive Science

Page 18: Mathematics Colloquium, UCSC

Scientific Tasks

Science Problem(IC, BC, ODE/PDE)

Simulator (code, computer)

Results(Validation, verification,

analysis)

Page 19: Mathematics Colloquium, UCSC

1. Mathematical Models

Hydrodynamics (gas dynamics)

@⇢

@t+r · (⇢v) = 0 mass eqn

@⇢v

@t+r · (⇢vv) +rP = ⇢g momentum eqn

@⇢E

@t+r · [(⇢E + P )v] = ⇢v · g total energy eqn

E = ✏+1

2|v|2

P = (� � 1)⇢✏@⇢✏

@t+r · [(⇢✏+ P )v]� v ·rP = 0

Equation of State

Page 20: Mathematics Colloquium, UCSC

2. Mathematical Models

Magnetohydrodynamics (MHD)@⇢

@t+r · (⇢v) = 0 mass eqn

@⇢v

@t+r · (⇢vv �BB) +rP⇤ = ⇢g +r · ⌧ momentum eqn

@⇢E

@t+r · [v(⇢E + P⇤)�B(v ·B)] = ⇢g · v +r · (v · ⌧ + �rT ) +r · (B⇥ (⌘r⇥B))

total energy eqn@B

@t+r · (vB�Bv) = �r⇥ (⌘r⇥B) induction eqn

P⇤ = p+B2

2

E =v2

2+ ✏+

B2

2⇢

Equation of State

⌧ = µ[(rv) + (rv)T � 2

3(r · v)I]

viscosity

r ·B = 0solenodidal constraint

Page 21: Mathematics Colloquium, UCSC

3. Mathematical Models

HEDP: Separate energy eqns for ion, electron, radiation (“3-temperature, or 3T”)

@

@t(⇢✏

ion

) +r · (⇢✏ion

v) + Pion

r · v = ⇢cv,ele⌧ei

(Tele

� Tion

) ion energy@

@t(⇢✏

ele

) +r · (⇢✏ele

v) + Pele

r · v = ⇢cv,ele⌧ei

(Tion

� Tele

)�r · qele

+Qabs

�Qemis

+Qlas

electron energy@

@t(⇢✏rad) +r · (⇢✏radv) + Pradr · v = r · qrad �Qabs +Qemis

radiation energy✏tot

= ✏ion

+ ✏ele

+ ✏rad

Ptot

, Tion

, Tele

, Trad

= EoS(⇢, ✏ion

, ✏ele

, ✏rad

) 3T EoS

Ptot

= EoS(⇢, ✏tot

)

@

@t(⇢✏

tot

) +r · (⇢✏tot

v) + Ptot

r · v = 0

Compare 3T with a simple 1T EoS!

Tion

= Tele

= Trad

, or Tele

= Tion

, Trad

= 0

Page 22: Mathematics Colloquium, UCSC

Discretization Approaches

‣Finite Volume (FV)

▪ shock capturing, compressible flows, structured/unstructured grids

▪ hard to achieve high-order (higher than 2nd order)

‣Finite Difference (FD)

▪ smooth flows, incompressible, high-order methods

▪ non-conservative, simple geometry

‣Finite Element (FE)

▪ arbitrary geometry, basis functions, continuous solutions

▪ hard coding, problems at strong gradients

‣Spectral Element (SE)

‣Discontinuous Galerkin (DG)

Page 23: Mathematics Colloquium, UCSC

Finite Volume Formulations

@U

@t

+@F

@x

+@G

@y

+@H

@z

= 0@U

@t+r · Flux(U) = 0

‣ Conservation laws (mass, momentum, energy)‣ Highly compressible flows with shocks and discontinuities‣ Differential (smooth) form of PDEs (e.g., FD) becomes invalid‣ Integral form of PDEs relaxes the smoothness assumptions and seek for weak solutions over control volumes and boundaries

‣ Basics of FV (in 1D):

x

tUn+1i

Uni Un

i+1Uni�1

tn+1

tn

Fni�1/2 Fn

i+1/2

or

Page 24: Mathematics Colloquium, UCSC

‣ Integral form of PDE:

‣ Volume averaged, cell-centered quantity & time averaged flux:

‣ Finite wave speed in hyperbolic system:

‣ General discrete difference equation in conservation form in 1D:

Finite Volume Formulations

Zxi+1/2

xi�1/2

u(x, tn+1)dx�

Zxi+1/2

xi�1/2

u(x, tn

)dx =

Ztn+1

tn

f(u(xi�1/2, t))dt�

Ztn+1

tn

f(u(xi+1/2, t))dt

U

n

i

=1

�x

Zxi+1/2

xi�1/2

u(x, tn

)dx F

ni�1/2 =

1

�t

Z tn+1

tn

f(u(xi�1/2, t))dtand

Fni�1/2 = F(Un

i�1, Uni )

* High-order reconstruction in space & time* Riemann problem at each cell-interface, i-1/2

U

n+1i = U

ni � �t

�x

(Fni+1/2 � F

ni�1/2)

Page 25: Mathematics Colloquium, UCSC

Riemann Problem & Godunov Method

‣The Riemann problem:

‣Two cases:

PDEs: Ut

+AU

x

= 0,�1 < x < 1, t > 0

IC : U(x, t = 0) = U

0(x) =

(U

L

if x < 0,

U

R

if x > 0.

Shock solution

Rarefaction solution

Page 26: Mathematics Colloquium, UCSC

A Discrete World of FV

U(x, tn)

xixi�1 xi+1

Page 27: Mathematics Colloquium, UCSC

A Discrete World of FV

u(xi, tn) = Pi(x), x 2 (xi�1/2, xi�1/2)

xixi�1 xi+1

piecewise polynomial reconstruction on each cell

uL = Pi+1(xi+1/2)uR = Pi(xi+1/2)

Page 28: Mathematics Colloquium, UCSC

A Discrete World of FV

xixi�1 xi+1

At each interface we solve a RP and obtain Fi+1/2

Page 29: Mathematics Colloquium, UCSC

A Discrete World of FV

We are ready to advance our solution in time and get new volume-averaged states

U

n+1i = U

ni � �t

�x

⇣F

⇤i+1/2 � F

⇤i�1/2

Page 30: Mathematics Colloquium, UCSC

PLM PPM

High-Order Polynomial Reconstruction

• Godunov’s order-barrier theorem (1959) • Monotonicity-preserving advection schemes are at most first-order! (Oh no…)• Only true for linear PDE theory (YES!)

• High-order “polynomial” schemes became available using non-linear slope limiters (70’s and 80’s: Boris, van Leer, Zalesak, Colella, Harten, Shu, Engquist, etc)

• Can’t avoid oscillations completely (non-TVD)• Instability grows (numerical INSTABILITY!)

FOG

Page 31: Mathematics Colloquium, UCSC

Low vs. High order Reconstructions

Page 32: Mathematics Colloquium, UCSC

Traditional High-Order Schemes

‣ Traditional approaches to get Nth high-order schemes take (N-1)th degree polynomial for interpolation/reconstruction

▪ only for normal direction (e.g., PLM, PPM, ENO, WENO, etc)

▪with monotonicity controls (e.g., slope limiters, artificial viscosity)

‣ High-order in FV is tricky (when compared to FD)

▪ volume-averaged quantities (quadrature rules)

▪ preserving conservation w/o losing accuracy

▪ higher the order, larger the stencil

▪ high-order temporal update (ODE solvers, e.g., RK3, RK4, etc.)

2D stencil for 2nd order PLM

2D stencil for 3rd order PPM

Page 33: Mathematics Colloquium, UCSC

Stability, Consistency and Convergence

‣Lax Equivalence Theorem (for linear problem, P. Lax, 1956)

▪ The only convergent schemes are those that are both consistent and stable.

▪Hard to show that the numerical solution converges to the original

solution of PDE; relatively easy to show consistency and stability of

numerical schemes

‣In practice, non-linear problems adopts the linear theory as guidance.

▪ code verification (code-to-code comparison)

▪ code validation (code-to-experiment, code-to-analytical solution

comparisons)

▪ self-convergence test over grid resolutions (a good measurement for

numerical accuracy)

Page 34: Mathematics Colloquium, UCSC

Shu-Osher Problem:1D Mach 3 Shock

Page 35: Mathematics Colloquium, UCSC

Various Reconstructions

Page 36: Mathematics Colloquium, UCSC

Circularly Polarized Alfven Wave

▪A CPAW problem propagates smoothly varying oscillations of the transverse components of velocity and magnetic field.

▪The initial condition is the exact nonlinear solutions of the MHD equations.

▪The decay of the max of Vz and Bz is solely due to numerical dissipation: direct measurement of numerical diffusion (Ryu, Jones & Frank, ApJ, 1995; Toth 2000, Del Zanna et al. 2001; Gardiner & Stone 2005, 2008).

These results are in agreement with previous investigations [5,6] and strongly supports the idea that problems involvingcomplex wave interactions may benefit from using higher-order schemes such as the ones presented here.

4.2. Shock tube problems

Shock tube problems are commonly used to test the ability of the scheme in describing both continuous and discontin-uous flow features. In the following we consider two- and three-dimensional rotated configurations of standard one-dimen-sional tubes. The default value for the parameter ap controlling monopole damping (see Eq. (9)) is 0.8.

4.2.1. Two-dimensional shock tubeFollowing [30,33], we consider a rotated version of the Brio-Wu test problem [8] with left and right states are given by

VL ¼ ð1; 0;0; 0:75;1;1ÞT for x1 < 0;

VR ¼ ð0:125;0;0;0:75;$1;0:1ÞT for x1 > 0;

(ð49Þ

where V ¼ ðq; v1;v2;B1;B2; pÞ is the vector of primitive variables. The subscript ‘‘1” gives the direction perpendicular to theinitial surface of discontinuity whereas ‘‘2” corresponds to the transverse direction. Here C = 2 is used and the evolution isinterrupted at time t = 0.2, before the fast waves reach the borders.

In order to address the ability to preserve the initial planar symmetry we rotate the initial condition by the angle a ¼ p=4in a two-dimensional plane with x 2 ½$1;1& and y 2 ½$0:01;0:01& using Nx ' Nx=100 grid points, with Nx = 600. Vectors followthe same transformation given by Eq. (45) with b = c = 0. This is known to minimize errors of the longitudinal component ofthe magnetic field (see for example the discussions in [52,27]). Boundary conditions respect the translational invariancespecified by the rotation: for each flow quantity we prescribe qði; jÞ ¼ qði( di; j( djÞ where ðdi; djÞ ¼ ð1;$1Þ, with the plus(minus) sign for the leftmost and upper (rightmost and lower) boundary. Computations are stopped before the fast rarefac-tion waves reach the boundaries, at t ¼ 0:2 cos a.

Fig. A.4 shows the primitive variable profiles for all schemes against a one-dimensional reference solution obtained on abase grid of 1024 zones with five levels of refinement. Errors in L1 norm, computed with respect to the same reference solu-tion, are sorted in Table 2 for density and the normal component of magnetic field. The out-coming wave pattern is com-prised, from left to right, of a fast rarefaction, a compound wave (an intermediate shock followed by a slow rarefaction),a contact discontinuity, a slow shock and a fast rarefaction wave. We see that all discontinuities are captured correctlyand the overall behavior matches the reference solution very well. The normal component of magnetic field is best describedwith MP5 and does not show erroneous jumps. Indeed, the profiles are essentially constant with small amplitude oscillationsshowing a relative peak )0.7%. Divergence errors, typically K 10$2, remain bounded with resolution and tend to saturate

Fig. A.3. Long term decay of circularly polarized Alfvén waves after 16.5 time units, corresponding to ) 100 wave periods. In the left panel, we plot themaximum value of the vertical component of velocity as a function of time for the WENO $ Z (solid line) and WENO + 3 (dashed line) schemes. Forcomparison, the dotted line gives the result obtained by a second-order TVD scheme. The panel on the right shows the analogous behavior of the verticalcomponent of magnetic field Bz for LimO3 and MP5. For all cases, the resolution was set to 120 ' 20 and the Courant number is 0.4.

A. Mignone et al. / Journal of Computational Physics 229 (2010) 5896–5920 5907

Source: Mignone & Tzeferacos, 2010, JCP

Page 37: Mathematics Colloquium, UCSC

Performance Comparison

L1 norm error

avg. comp. time / step

32 256

0.221 (x5/3)sec 38.4 sec

Source: Mignone & Tzeferacos, 2010, JCP

▪PPM (overall 2nd order): 2h42m50s

▪MP5 (5th order): 15s(x5/3)=25s

▪More computational work & less memory

▪ Better suited for HPC

▪ Easier in FD; harder in FV

▪High-orders schemes are better in preserving solution accuracy on AMR.

Page 38: Mathematics Colloquium, UCSC

Truncation Errors at Fine-Coarse Boundary

✓Any 2nd order Scheme becomes 1st order at fine-coarse boundaries.

✓The deeper AMR level, the worse truncation errors accumulated and solutions will become 1st order almost everywhere if grid pattern changes frequently.

✓High-order scheme is NOT just an option! (see papers by Colella et al.)

F f,Li�1/2,j+1/4

F f,Li�1/2,j�1/4

F c,Ri�1/2,j(i, j)

�U

�t+r · F = TE =

(O(�h) at F/C boundary

O(�h2) otherwise

DF =1

�x

hF c,Ri+1/2,j �

1

2(F f,L

i�1/2,j+1/4 + F f,Li�1/2,j�1/4)

i

=1

�x

hF�(i+ 1/2)�x, j�y

�� F

�(i� 1/2)�x, j�y

�+O(�y2)

i

=@F

@x+O(�h), assuming �x ⇡ �y(= �h)

Page 39: Mathematics Colloquium, UCSC

Multidimensional Formulation

‣ 2D discrete difference equation in conservation form:

‣ Two different approaches:‣ directionally “split” formulation‣ update each spatial direction separately, easy to implement, robust‣ always good?

‣ directionally “unsplit” formulation‣ update both spatial directions at the same time, harder to implement‣ you gain improved extra bonus (i.e., stability) from what you pay for!

U

n+1i,j = U

ni,j �

�t

�x

(Fni+1/2,j � F

ni�1/2,j)�

�t

�y

(Gni,j+1/2 �G

ni,j�1/2)

Page 40: Mathematics Colloquium, UCSC

Unsplit FV Formulation

‣ 2D discrete difference equation in conservation form:

U

n+1i,j = U

ni,j �

u�t

�x

[Uni,j � U

ni�1,j ]�

v�t

�y

[Uni,j � U

ni,j�1] U

n+1i,j = U

ni,j �

u�t

�x

[Uni,j � U

ni�1,j ]�

v�t

�y

[Uni,j � U

ni,j�1]

+�t

2

2

n

u

�x

v

�y

(Uni,j � U

ni,j�1)�

v

�y

(Uni�1,j � U

ni�1,j�1)

+v

�y

v

�y

(Uni,j � U

ni�1,j)�

v

�y

(Uni,j�1 � U

ni�1,j�1)

o

Extra cost for corner coupling!

(a) 1st order donor cell

(i, j)

U

n+1i,j = U

ni,j �

�t

�x

(Fni+1/2,j � F

ni�1/2,j)�

�t

�y

(Gni,j+1/2 �G

ni,j�1/2)

(b) 2nd order corner-transport-upwind (CTU)

(i, j)

Page 41: Mathematics Colloquium, UCSC

Unsplit FV Formulation

‣ 2D discrete difference equation in conservation form:

U

n+1i,j = U

ni,j �

�t

�x

(Fni+1/2,j � F

ni�1/2,j)�

�t

�y

(Gni,j+1/2 �G

ni,j�1/2)

����u�t

�x

����+����v�t

�y

���� 1max

⇣����u�t

�x

���� ,����v�t

�y

����⌘ 1

Smaller stability region Gain: Extended stability region

(a) 1st order donor cell

(i, j)

(b) 2nd order corner-transport-upwind (CTU)

(i, j)

Page 42: Mathematics Colloquium, UCSC

Unsplit vs. Split

✓Single-mode RT instability (Almgren et al. ApJ, 2010)

✓Split solver:High-wavenumber instabilities grow due to experiencing high compression and expansion in each directional sweep.

✓Unsplit solver:High-wavenumber instabilities are suppressed and do not grow.

✓ For MHD, it is more crucial to use unsplit in order to preserve divergence-free solenoidal constraint (Lee & Deane, 2009; Lee, 2013):

Split PPM

Unsplit PPM

r ·B =@B

x

@x

+@B

y

@y

+@B

z

@z

= 0

Page 43: Mathematics Colloquium, UCSC

Unsplit vs. Split

unsplit PPM split PPM

Page 44: Mathematics Colloquium, UCSC

Excellent Resources

Page 45: Mathematics Colloquium, UCSC

Last Episode

1. High Performance Computing

2. Ideas on Numerical Methods

3. Validation & Predictive Science

Page 46: Mathematics Colloquium, UCSC

In Collaboration with

U of Chicago: D. Q. Lamb

P. TzeferacosN. Flocke

C. GrazianiK. Weide

U of Oxford:G. GregoriJ. Meinecke

Page 47: Mathematics Colloquium, UCSC

To Investigate B-field in the Universe

• In the universe, shocks are driven when two or more giant galaxy clusters are merging together by gravitational collapsing.

• Mass accretion onto these clusters generate high Mach number shocks.

• These shocks can form a tiny “seed” magnetic fields which can then be amplified by turbulent dynamo processes.

Courtesy of F. Miniati (ETH)

Filaments

ClustersVoids

Expanding Shocks

Courtesy of A. KravtsovU of Chicago

Page 48: Mathematics Colloquium, UCSC

Shock Waves In SNR

Inner Region• The interior of the SNR Cassiopeia A contains a disordered shell of radio synchrotron emission by giga-electronvolt electrons.

•The inferred magnetic field in these radio knots is a few milli-gauss, about 100 times larger than the surrounding interstellar gas.

ForwardShock Wave

ShockedEjectadue to

Reverse Shocks

CircumstellarGas

CoolEjecta

Outer Region• Narrow X-ray filaments (~ parsec in width) at the outer shock rim are produced by synchrotron radiation from ultra-relativistic electrons.

• The B-fields in the outer region ~ 100µG or more.

Page 49: Mathematics Colloquium, UCSC

Biermann Battery Mechanism

• The origin of the magnetic fields in the universe is still not yet fully understood. • Generalized Ohm’s law:

J = r⇥B, where

(1) dynamo term

(2) resistive term

(3) Hall term

(4) Biermann battery term (BBT)

E =� u⇥ B

c+ ⌘J

+J⇥ B

cnee

� rpenee

The BBT is the mostly widely invoked

mechanism to produce B fields from

unmagnetized plasma condition

Page 50: Mathematics Colloquium, UCSC

Takeaway

@B

@t

���BBT

= crPe ⇥rne

qen2e

• BBT generates B fields when gradients of electron pressure and density are not aligned.

• BBT is zero in 1D (or symmetric) flow or at spherical shocks.

• BBT becomes non-zero when symmetry is broken, and two gradients are not aligned to each other, that is, at downstream of asymmetric shock. BBT = 0 BBT≠0BBT = 0

We use laser to drive asymmetric shocks! Please see our new study on BBT: Graziani et al., submitted to ApJ, 2014

Page 51: Mathematics Colloquium, UCSC

Magnetic Fields in HEDP

Tzeferacos et al., HEDP, 2014, Accepted

Meinecke et al., Nature Physics, 2014

Page 52: Mathematics Colloquium, UCSC

3cm$300$J$1ns$$2ω$

3,axis$coils$

Ar$(1atm)$

4cm$

Carbon$rod$

4cm$

Investigating B-field in the Universe

In collaboration with the research teams in

U of Chicago & Oxford Univ.Experimental Configuration

Page 53: Mathematics Colloquium, UCSC

In collaboration with the research teams in

U of Chicago & Oxford Univ.

Investigating B-field in the Universe

3D Simulation

Page 54: Mathematics Colloquium, UCSC

Validation & Predictive Science

breakout)shock)

material)discon2nuity)

shock)

shock)generated)B"

laser5target)B"

t="50ns"

Laser&e

nergy&range&used

&to&calibrate&ε&

• We used five Vulcan laser shots (280-300J) to calibrate the fraction of the laser energy deposited in the carbon rod target.

• With the calibrated amount of laser energy deposition we predicted the shock radius of six laser shots with energies ranging from 200-343J.

Page 55: Mathematics Colloquium, UCSC

Validation & Predictive Science

Experimental,data,FLASH,

shock,radius,(m

m),

t,(μs),

Meinecke et al., Nature Physics, 2-14

Page 56: Mathematics Colloquium, UCSC

Summary

▪Novel ideas of numerical algorithms can play the key fundamental

role in many areas in modern science.

▪Computational mathematics is one of the corner stone research

areas that can provide major predictive scientific tools.

▪Numerical simulations can help designing better scientific

directions in wide ranges of research applications, especially in

physical science and engineering.

Page 57: Mathematics Colloquium, UCSC

Thanks! Questions?

Dongwook Lee: [email protected]

ams.soe.ucsc.edu/people/dongwook

FLASH: www.flash.uchicago.edu

Page 58: Mathematics Colloquium, UCSC

Supplementary Slides

Page 59: Mathematics Colloquium, UCSC

Relavant Questions

‣ What is scientific computing?

‣ How do we want to use computer for it?

‣ What should we do in order to use computing resources in a better way?

‣ Can numerical algorithms and computational mathematics improve computations?

‣ High-performance computing: petascale (or exascale ?)

Page 60: Mathematics Colloquium, UCSC

Various Reconstruction

1st

2nd

3rd

5th

200 cells

1st on 400 cells

1st on 800 cells

1st + HLL

200 cells

1st + HLLC

1st + Roe

Page 61: Mathematics Colloquium, UCSC

Low-Order vs. High-Order

1st OrderHigh-Order

Ref. Soln

1st order: 3200 cells (50 MB), 160 sec, 3828 stepsvs.

High-order: 200 cells (10 MB), 9 sec, 266 steps