PDR.PODM Distributed Memory

From crtc.cs.odu.edu
Revision as of 01:07, 24 March 2023 by Pthomadakis (talk | contribs) (Latest Results)
Jump to: navigation, search

Issues

  • No reuse of leaves refined by worker nodes. The picture below shows the issue. Two neighbour leaves (0,1) each refined as the main leaf (0 top, 1 bottom) but not refined as a neighor.

PDR PODM Leaves not refined.png

  • Current algorithm uses neighbour traversal to distribute cells to octree leaving some cells out in some cases. Such a case can happen when a cell is part of an octree leaf based on its circumcenter but

it does not have any neighbour in the same leaf.

PDR PODM Cells not distributed.png

  • During unpacking the incident cell for each vertex is not set correctly. Specifically, in the case that the initial incident cell is not part of the working unit (Leaf + LVL.1 Neighbours) and thus is not local,

it is set to the infinite cell. This causes PODM to crash randomly for some cases.

  • Another issue comes from the way global IDs are updated for each cell's neighbors' IDs. The code that updates the cell's connectivity using global IDs takes the neighbor's pointer, retrieves its global ID and

updates the neighborID field. However, when the neighbour is part of another work unit's leaf and is not local this pointer is NULL. In this case the neighborID field is wrongly reset to the infinite cell ID, which as result, deletes the connectivity information forever.

  • The function that unpacks the required leaves before refinement does not discard duplicate vertices. Duplicate vertices will always be present since each leaf is packed and sent individually, and as a result,

neighbouring leaves will include the shared vertices. Because duplicate vertices are not handled, multiple vertex objects are created that are in fact the same point geometrically. Thus, two cells that share a common vertex could have pointers to two different vertex objects and, as a result, each cell views a different state about the same vertex.

Fixes

PDR Fix.png

Work Unit After Refinement.png


Interesting Findings

Delta 0.880

15 MPI ranks depth: 3

15 MPI ranks depth: 4

40 MPI ranks / 40 cores depth: 4

[Expand]

160 MPI ranks / 10 cores depth: 4

[Expand]

After Parallel int to pointer

15 MPI ranks depth: 3

[Expand]

15 MPI ranks depth: 4

[Expand]

40 MPI ranks depth: 4

[Expand]

160 MPI ranks 10 cores depth: 4

[Expand]

After Parallel int to pointer and leaf distribution

15 MPI ranks depth: 4

[Expand]

40 MPI ranks depth: 4

[Expand]

160 MPI ranks 10 cores depth: 4

[Expand]

After Parallel int to pointer, leaf distribution and extra comm thread

15 MPI ranks depth: 3

[Expand]

15 MPI ranks depth: 4

[Expand]

40 MPI ranks depth: 4

[Expand]

160 MPI ranks depth: 4

[Expand]

Delta 0.3780

15 MPI ranks depth: 4


Latest Results

Shared MemoryTimes
Cores Total Time (s) Number of Elements
40 107.8 216099955


MPI Times
MPI PREMA
Cores Total Time (s) Number of Elements Total Time (s) Number of Elements
100 1151.472406 49352359 208.746450 49347855
200 763.70671 49357898 121.326012 49353442
300 537.678638 49357092 105.812248 49351224
400 490.365970 49357881 93.101481 49351626
500 434.921334 49347357 93.119187 49351049
600 466.822803 49361647 96.243807 49345030
700 425.273205 49360615 97.721654 49346798
800 434.638603 49349629 96.723318 49341034

Pdr.mpi.png