File size: 6,799 Bytes
f6cc031
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
<Poster Width="1734" Height="1060">
	<Panel left="16" right="170" width="547" height="205">
		<Text>Abstract</Text>
		<Text>We introduce a new approach to automatically extract an idealized logical</Text>
		<Text>structure from a parallel execution trace. We use this structure to define</Text>
		<Text>intuitive metrics such as the lateness of a process involved in a parallel</Text>
		<Text>execution. By analyzing and illustrating traces in terms of logical steps, we</Text>
		<Text>leverage a developer’s understanding of the happened-before relations in</Text>
		<Text>a parallel program. This technique can uncover dependency chains,</Text>
		<Text>elucidate communication patterns, and highlight sources and propagation</Text>
		<Text>of delays, all of which may be obscured in a traditional trace visualization.</Text>
	</Panel>

	<Panel left="13" right="387" width="548" height="623">
		<Text>Extracting Logical Structure</Text>
		<Text>The logical structure of a program is the ordering of events implied by that</Text>
		<Text>program. We describe the logical structure by assigning a logical step to</Text>
		<Text>each event.</Text>
		<Text>Structure extraction occurs in two phases:</Text>
		<Text>1.  Partitioning related communication</Text>
		<Text>2.  Step assignment</Text>
		<Text>Partitioning</Text>
		<Text>Partitions represent non-overlapping application phases. If not predefined,</Text>
		<Text>we derive them from the trace:</Text>
		<Text>Matching sends and receives and communication handled by the same</Text>
		<Text>MPI call must be related and thus in the same partition. When merged,</Text>
		<Text>this can create cycles in ordering:</Text>
		<Figure left="24" right="711" width="263" height="112" no="1" OriWidth="0" OriHeight="0
" />
		<Figure left="309" right="705" width="250" height="105" no="2" OriWidth="0" OriHeight="0
" />
		<Text>Communication partitions forming a cycle do not permit a partial order, so</Text>
		<Text>we infer these partitions are related and merge them.</Text>
		<Text>In addition to merging due to ordering</Text>
		<Text>constraints, we can optionally merge due</Text>
		<Text>to behavioral assumptions. For example,</Text>
		<Text>in bulk synchronous codes we expect</Text>
		<Text>each process to be active at some</Text>
		<Text>distance in the partition graph.</Text>
		<Figure left="342" right="867" width="215" height="129" no="3" OriWidth="0" OriHeight="0
" />
	</Panel>

	<Panel left="595" right="171" width="545" height="369">
		<Text>Step Assignment</Text>
		<Text>Each partition is independently assigned steps based on two principles:</Text>
		<Text>1.  Happened-before relationships must be maintained</Text>
		<Text>2.  Send events have greater impact on structure</Text>
		<Figure left="602" right="268" width="228" height="69" no="4" OriWidth="0" OriHeight="0
" />
		<Text>Consider this trace segment from an 8-</Text>
		<Text>process run of the pF3D stencil</Text>
		<Text>communication benchmark [1].</Text>
		<Text>First we determine groups of</Text>
		<Text>simultaneous sends (gray) using</Text>
		<Text>receives only for ordering.</Text>
		<Figure left="899" right="343" width="236" height="69" no="5" OriWidth="0" OriHeight="0
" />
		<Figure left="599" right="416" width="240" height="70" no="6" OriWidth="0" OriHeight="0
" />
		<Text>Then we assign the least step</Text>
		<Text>possible to each event.</Text>
		<Text>Finally we insert aggregated non-communication events between the</Text>
		<Text>sends and receives and determine global steps using partition ordering.</Text>
	</Panel>

	<Panel left="594" right="558" width="546" height="453">
		<Text>Temporal Metrics</Text>
		<Text>Having determined a logical structure, we can calculate how late an event</Text>
		<Text>was relative to its peers. We define lateness as excess completion time</Text>
		<Text>over the earliest related event at a step.</Text>
		<Text>We visualize a portion of an MG [2] trace using traditional methods as</Text>
		<Text>represented by Vampir [3] (left) and logical structure and lateness (right).</Text>
		<Text>In the latter the communication pattern and delay propagation is clear.</Text>
		<Figure left="602" right="724" width="258" height="108" no="7" OriWidth="0.184544" OriHeight="0.0597148
" />
		<Figure left="872" right="726" width="260" height="106" no="8" OriWidth="0.191465" OriHeight="0.0610517
" />
		<Text>We classify four situations contributing to event lateness:</Text>
		<Figure left="595" right="866" width="540" height="54" no="9" OriWidth="0.398501" OriHeight="0.0534759
" />
		<Text>Using this classification, we can narrow our focus to events where</Text>
		<Text>lateness originates by subtracting out propagated lateness. This</Text>
		<Text>differential lateness allows us to pinpoint sources of delays automatically.</Text>
	</Panel>

	<Panel left="1171" right="168" width="550" height="731">
		<Text>Case Study</Text>
		<Text>We analyze a massively parallel algorithm to compute merge trees. The</Text>
		<Text>algorithm relies on a global gather-scatter approach where each level</Text>
		<Text>requires messages sent both up and down a k-ary gather tree:</Text>
		<Figure left="1177" right="262" width="538" height="85" no="10" OriWidth="0" OriHeight="0
" />
		<Text>Below are the Vampir (left) and logical structure (right) visualizations of a</Text>
		<Text>16 process, 4-ary merge tree calculation. In the logical structure view,</Text>
		<Text>lateness reflects data-dependent load imbalance. Logical steps highlight</Text>
		<Text>the gather tree structure, revealing that the gather processes send back to</Text>
		<Text>the leaves before sending up to the root, missing an opportunity for more</Text>
		<Text>aggressive pipelining.</Text>
		<Figure left="1179" right="471" width="276" height="71" no="11" OriWidth="0.392157" OriHeight="0.0757576
" />
		<Figure left="1466" right="475" width="247" height="65" no="12" OriWidth="0.390427" OriHeight="0.0815508
" />
		<Text>The 1024-process, 8-ary tree below shows similar issues. The recurring</Text>
		<Text>“panhandle” shape highlights waiting due to sending down before up.</Text>
		<Figure left="1180" right="602" width="534" height="290" no="13" OriWidth="0" OriHeight="0
" />
	</Panel>

	<Panel left="1173" right="915" width="546" height="97">
		<Text>References</Text>
		<Text>1.  C. H. Still et al. Filamentation nd forward brillouin scatter of entire smoothed and aberrated laser beams.</Text>
		<Text>Physics of Plasmas, 7(5):2023, 2000.</Text>
		<Text>2.  D. H. Bailey et al. The nas parallel benchmarks. Int. J. Supercomput. Appl., 5(3):63–73, 1991.</Text>
		<Text>3.  W. E. Nagel et al. VAMPIR: Visualization and analysis of MPI resources. Supercomputer, 12(1):69-80, 1996.</Text>
	</Panel>

</Poster>