Computational cost of isogeometric multi-frontal solvers on parallel distributed memory machines

This paper derives theoretical estimates of the computational cost for isogeometric multi-frontal direct solver executed on parallel distributed memory machines. We show theoretically that for the Cp-1 global continuity of the isogeometric solution, both the computational cost and the communication...

Full description

Bibliographic Details
Main Authors: Wozniak, M., Paszynski, M., Pardo, D., Dalcin, L., Calo, Victor
Format: Journal Article
Published: 2015
Online Access:http://hdl.handle.net/20.500.11937/51579
_version_ 1848758732984418304
author Wozniak, M.
Paszynski, M.
Pardo, D.
Dalcin, L.
Calo, Victor
author_facet Wozniak, M.
Paszynski, M.
Pardo, D.
Dalcin, L.
Calo, Victor
author_sort Wozniak, M.
building Curtin Institutional Repository
collection Online Access
description This paper derives theoretical estimates of the computational cost for isogeometric multi-frontal direct solver executed on parallel distributed memory machines. We show theoretically that for the Cp-1 global continuity of the isogeometric solution, both the computational cost and the communication cost of a direct solver are of order O(log(N)p2) for the one dimensional (1D) case, O(Np2) for the two dimensional (2D) case, and O(N4/3p2) for the three dimensional (3D) case, where N is the number of degrees of freedom and p is the polynomial order of the B-spline basis functions. The theoretical estimates are verified by numerical experiments performed with three parallel multi-frontal direct solvers: MUMPS, PaStiX and SuperLU, available through PETIGA toolkit built on top of PETSc. Numerical results confirm these theoretical estimates both in terms of p and N. For a given problem size, the strong efficiency rapidly decreases as the number of processors increases, becoming about 20% for 256 processors for a 3D example with 1283 unknowns and linear B-splines with C0 global continuity, and 15% for a 3D example with 643 unknowns and quartic B-splines with C3 global continuity. At the same time, one cannot arbitrarily increase the problem size, since the memory required by higher order continuity spaces is large, quickly consuming all the available memory resources even in the parallel distributed memory version. Numerical results also suggest that the use of distributed parallel machines is highly beneficial when solving higher order continuity spaces, although the number of processors that one can efficiently employ is somehow limited.
first_indexed 2025-11-14T09:48:40Z
format Journal Article
id curtin-20.500.11937-51579
institution Curtin University Malaysia
institution_category Local University
last_indexed 2025-11-14T09:48:40Z
publishDate 2015
recordtype eprints
repository_type Digital Repository
spelling curtin-20.500.11937-515792018-02-20T01:01:42Z Computational cost of isogeometric multi-frontal solvers on parallel distributed memory machines Wozniak, M. Paszynski, M. Pardo, D. Dalcin, L. Calo, Victor This paper derives theoretical estimates of the computational cost for isogeometric multi-frontal direct solver executed on parallel distributed memory machines. We show theoretically that for the Cp-1 global continuity of the isogeometric solution, both the computational cost and the communication cost of a direct solver are of order O(log(N)p2) for the one dimensional (1D) case, O(Np2) for the two dimensional (2D) case, and O(N4/3p2) for the three dimensional (3D) case, where N is the number of degrees of freedom and p is the polynomial order of the B-spline basis functions. The theoretical estimates are verified by numerical experiments performed with three parallel multi-frontal direct solvers: MUMPS, PaStiX and SuperLU, available through PETIGA toolkit built on top of PETSc. Numerical results confirm these theoretical estimates both in terms of p and N. For a given problem size, the strong efficiency rapidly decreases as the number of processors increases, becoming about 20% for 256 processors for a 3D example with 1283 unknowns and linear B-splines with C0 global continuity, and 15% for a 3D example with 643 unknowns and quartic B-splines with C3 global continuity. At the same time, one cannot arbitrarily increase the problem size, since the memory required by higher order continuity spaces is large, quickly consuming all the available memory resources even in the parallel distributed memory version. Numerical results also suggest that the use of distributed parallel machines is highly beneficial when solving higher order continuity spaces, although the number of processors that one can efficiently employ is somehow limited. 2015 Journal Article http://hdl.handle.net/20.500.11937/51579 10.1016/j.cma.2014.11.020 fulltext
spellingShingle Wozniak, M.
Paszynski, M.
Pardo, D.
Dalcin, L.
Calo, Victor
Computational cost of isogeometric multi-frontal solvers on parallel distributed memory machines
title Computational cost of isogeometric multi-frontal solvers on parallel distributed memory machines
title_full Computational cost of isogeometric multi-frontal solvers on parallel distributed memory machines
title_fullStr Computational cost of isogeometric multi-frontal solvers on parallel distributed memory machines
title_full_unstemmed Computational cost of isogeometric multi-frontal solvers on parallel distributed memory machines
title_short Computational cost of isogeometric multi-frontal solvers on parallel distributed memory machines
title_sort computational cost of isogeometric multi-frontal solvers on parallel distributed memory machines
url http://hdl.handle.net/20.500.11937/51579