TY - JOUR
T1 - Simulating radiating and magnetized flows in multiple dimensions with ZEUS-MP
AU - Hayes, John C.
AU - Norman, Michael L.
AU - Fiedler, Robert A.
AU - Bordner, James O.
AU - Li, Pak Shing
AU - Clark, Stephen E.
AU - Ud-Doula, Asif
AU - Low, Mordecai Mark Mac
PY - 2006/7
Y1 - 2006/7
N2 - This paper describes ZEUS-MP, a multiphysics, massively parallel, message-passing implementation of the ZEUS code. ZEUS-MP differs significantly from the thoroughly documented ZEUS-2D code, the completely undocumented (in peer-reviewed literature) ZEUS-3D code, and a marginally documented "version 1" of ZEUS-MP first distributed publicly in 1999. ZEUS-MP offers an MHD algorithm that is better suited for multidimensional flows than the ZEUS-2D module by virtue of modifications to the method of characteristics scheme first suggested by Hawley & Stone. This MHD module is shown to compare quite favorably to the TVD scheme described by Ryu et al. ZEUS-MP is the first publicly available ZEUS code to allow the advection of multiple chemical (or nuclear) species. Radiation hydrodynamic simulations are enabled via an implicit flux-limited radiation diffusion (FLD) module. The hydrodynamic, MHD, and FLD modules can be used, singly or in concert, in one, two, or three space dimensions. In addition, so-called 1.5D and 2.5D grids, in which the "half-D" denotes a symmetry axis along which a constant but nonzero value of velocity or magnetic field is evolved, are supported. Self-gravity can be included either through the assumption of a GM/r potential or through a solution of Poisson's equation using one of three linear solver packages (conjugate gradient, multigrid, and FFT) provided for that purpose. Point-mass potentials are also supported. Because ZEUS-MP is designed for large simulations on parallel computing platforms, considerable attention is paid to the parallel performance characteristics of each module in the code. Strong-scaling tests involving pure hydrodynamics (with and without self-gravity), MHD, and RHD are performed in which large problems (256 3 zones) are distributed among as many as 1024 processors of an IBM SP3. Parallel efficiency is a strong function of the amount of communication required between processors in a given algorithm, but all modules are shown to scale well on up to 1024 processors for the chosen fixed problem size.
AB - This paper describes ZEUS-MP, a multiphysics, massively parallel, message-passing implementation of the ZEUS code. ZEUS-MP differs significantly from the thoroughly documented ZEUS-2D code, the completely undocumented (in peer-reviewed literature) ZEUS-3D code, and a marginally documented "version 1" of ZEUS-MP first distributed publicly in 1999. ZEUS-MP offers an MHD algorithm that is better suited for multidimensional flows than the ZEUS-2D module by virtue of modifications to the method of characteristics scheme first suggested by Hawley & Stone. This MHD module is shown to compare quite favorably to the TVD scheme described by Ryu et al. ZEUS-MP is the first publicly available ZEUS code to allow the advection of multiple chemical (or nuclear) species. Radiation hydrodynamic simulations are enabled via an implicit flux-limited radiation diffusion (FLD) module. The hydrodynamic, MHD, and FLD modules can be used, singly or in concert, in one, two, or three space dimensions. In addition, so-called 1.5D and 2.5D grids, in which the "half-D" denotes a symmetry axis along which a constant but nonzero value of velocity or magnetic field is evolved, are supported. Self-gravity can be included either through the assumption of a GM/r potential or through a solution of Poisson's equation using one of three linear solver packages (conjugate gradient, multigrid, and FFT) provided for that purpose. Point-mass potentials are also supported. Because ZEUS-MP is designed for large simulations on parallel computing platforms, considerable attention is paid to the parallel performance characteristics of each module in the code. Strong-scaling tests involving pure hydrodynamics (with and without self-gravity), MHD, and RHD are performed in which large problems (256 3 zones) are distributed among as many as 1024 processors of an IBM SP3. Parallel efficiency is a strong function of the amount of communication required between processors in a given algorithm, but all modules are shown to scale well on up to 1024 processors for the chosen fixed problem size.
UR - http://www.scopus.com/inward/record.url?scp=33748521208&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=33748521208&partnerID=8YFLogxK
U2 - 10.1086/504594
DO - 10.1086/504594
M3 - Article
AN - SCOPUS:33748521208
SN - 0067-0049
VL - 165
SP - 188
EP - 228
JO - Astrophysical Journal, Supplement Series
JF - Astrophysical Journal, Supplement Series
IS - 1
ER -