Cores

From crtc.cs.odu.edu
Revision as of 11:56, 19 August 2022 by Nikos (talk | contribs)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to: navigation, search

This page lists some of the past activities within the different cores for metrics-driven extreme-scale parallel anisotropic mesh generation. The current focus is on quantum computing and the core thrusts will be posted soon.


Parallel Adaptive Mesh Generation

Introduction

  • Telescopic.png

    Our long term goal is to achieve extreme-scale adaptive CFD simulations on the complex, heterogeneous High Performance Computing (HPC) platforms. To achieve this goal, we proposed in a telescopic approach (see Figure) [1]. The telescopic approach is critical in leveraging the concurrency that exists at multiple levels in anisotropic and adaptive simulations. At the chip and node levels, the telescopic approach deploys a Parallel Optimistic (PO) layer and Parallel Data Refinement (PDR) layer, respectively. Finally, between different nodes the loosely-coupled approach is employed

    The requirements for parallel mesh generation and adaptivity are:

    1. Stability: The quality of the mesh generated in parallel must be comparable to that of a mesh generated sequentially. The quality is defined in terms of the shape of the elements (using a chosen location/error-dependent metric), and the number of the elements (fewer is better for the same shape constraint).
    2. Robustness: the ability of the software to correctly and efficiently process any input data. Operator intervention into a massively parallel computation is not only highly expensive, but most likely infeasible due to the large number of concurrently processed sub-problems.
    3. Code re-use: a modular design of the parallel software that builds upon previously designed sequential meshing code, such that it can be replaced and/or updated with a minimal effort. Code re-use is feasible only if the code satisfies the reproducibility criterion, identified for the first time in this project. However, the experience from this project indicated that the complexity of state-of-the-art codes inhibits their modifications to a degree that their integration with parallel frameworks like PDR becomes impractical.
    4. Scalability: the ratio of the time taken by the best sequential implementation to the time taken by the parallel implementation. The speedup is always limited by the inverse of the sequential fraction of the software, and therefore all non-trivial stages of the computation must be parallelized to leverage the current architectures with millions of cores.
    5. Reproducibility: Which can be expressed into two forms:
    1. Strong Reproducibility requires that the sequential mesh generation code, when executed with the same input, produces identical results under the following modes of execution:
      1. continuous without restarts,and
      2. with restarts and reconstructions of the internal data structures.
    2. Weak Reproducibility requires that the sequential mesh generation code, when executed with the same input, produces results of the same quality under the following modes of execution:
      1. continuous without restarts, and
      2. with restarts and reconstructions of the internal data structures.

  • Tightly-coupled approaches

    This groups is experimented with two tightly-coupled approaches :

    1. CDT3D
    2. Parallel Optimistic Delaunay Mesh Generation for medical images, for details see [2]

    Partially-coupled approaches

    Data Decomposition
    1. Parallel Data Refinement using TetGen
    2. Parallel Data Refinement using AFLR
    3. Parallel Data Refinement using Parallel Optimistic Delaunay Mesh Generation for medical images , for details see [3] and [4]


    Domain Decomposition
    PCDM
    extension to 3D is in progress.

    Loosely-coupled approach

    Loosely coupled using CDT3D

    Parallel Runtime Software Systems

    Real-time Medical Image Computing

    Real Time Medical Image Computing