68
Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text “Introduction to Parallel Computing”, Addison Wesley, 2003.

Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Embed Size (px)

Citation preview

Page 1: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Sorting Algorithms

Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar

To accompany the text “Introduction to Parallel Computing”,Addison Wesley, 2003.

Page 2: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Topic Overview

• Issues in Sorting on Parallel Computers

• Sorting Networks

• Bubble Sort and its Variants

• Quicksort

• Bucket and Sample Sort

• Other Sorting Algorithms

– Typeset by FoilTEX – 1

Page 3: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Sorting: Overview

• One of the most commonly used and well-studied kernels.

• Sorting can be comparison-based or noncomparison-based.

• The fundamental operation of comparison-based sorting iscompare-exchange.

• The lower bound on any comparison-based sort of n numbersis Θ(n log n).

• We focus here on comparison-based sorting algorithms.

– Typeset by FoilTEX – 2

Page 4: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Sorting: Basics

What is a parallel sorted sequence? Where are the input andoutput lists stored?

• We assume that the input and output lists are distributed.

• The sorted list is partitioned with the property that eachpartitioned list is sorted and each element in processor Pi’s listis less than that in Pj’s list if i < j.

– Typeset by FoilTEX – 3

Page 5: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Sorting: Parallel Compare Exchange Operation

Step 1 Step 2 Step 3

PSfrag replacements

ai aj

Pi PiPi PjPj Pj

maxai, ajminai, ajaj, aiai, aj

A parallel compare-exchange operation. Processes Pi and Pj

send their elements to each other. Process Pi keeps minai, aj,and Pj keeps maxai, aj.

– Typeset by FoilTEX – 4

Page 6: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Sorting: Basics

What is the parallel counterpart to a sequential comparator?

• If each processor has one element, the compare exchangeoperation stores the smaller element at the processor withsmaller id. This can be done in ts + tw time.

• If we have more than one element per processor, we call thisoperation a compare split. Assume each of two processorshave n/p elements.

• After the compare-split operation, the smaller n/p elements areat processor Pi and the larger n/p elements at Pj, where i < j.

• The time for a compare-split operation is (ts + twn/p), assumingthat the two partial lists were initially sorted.

– Typeset by FoilTEX – 5

Page 7: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Sorting: Parallel Compare Split Operation

861 11 13

1 2 6 7 82 6 9 12 13

1311

1 6 9 12 137 8 1110 10

11861 13 2 97 1210

87

1 86

211 1

2 97 12

2 97 12

10

10

Step 2

9 12 131110

Step 4Step 3

Step 1

PSfrag replacements

Pi

Pi Pi

PiPj

Pj Pj

Pj

A compare-split operation. Each process sends its block of sizen/p to the other process. Each process merges the received

block with its own block and retains only the appropriate half ofthe merged block. In this example, process Pi retains the smaller

elements and process Pj retains the larger elements.

– Typeset by FoilTEX – 6

Page 8: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Sorting Networks

• Networks of comparators designed specifically for sorting.

• A comparator is a device with two inputs x and y and twooutputs x′ and y′. For an increasing comparator , x′ = minx, yand y′ = maxx, y; and vice-versa.

• We denote an increasing comparator by ⊕ and a decreasingcomparator by .

• The speed of the network is proportional to its depth.

– Typeset by FoilTEX – 7

Page 9: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Sorting Networks: Comparators

(a)

(b)

PSfrag replacements

x

x

x

x

y

y

y

y

x′ = minx, y

y′ = maxx, y

x′ = maxx, y

y′ = minx, y

x′ = minx, y

y′ = maxx, y

x′ = maxx, y

y′ = minx, y

A schematic representation of comparators: (a) an increasingcomparator, and (b) a decreasing comparator.

– Typeset by FoilTEX – 8

Page 10: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Sorting NetworksColumns of comparators

Inpu

t wir

es

Out

put w

ires

Inte

rcon

nect

ion

netw

ork

A typical sorting network. Every sorting network is made up of aseries of columns, and each column contains a number of

comparators connected in parallel.

– Typeset by FoilTEX – 9

Page 11: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Sorting Networks: Bitonic Sort

• A bitonic sorting network sorts n elements in Θ(log2 n) time.

• A bitonic sequence has two tones – increasing anddecreasing, or vice versa. Any cyclic rotation of such networksis also considered bitonic.

• 〈1, 2, 4, 7, 6, 0〉 is a bitonic sequence, because it first increasesand then decreases. 〈8, 9, 2, 1, 0, 4〉 is another bitonic sequence,because it is a cyclic shift of 〈0, 4, 8, 9, 2, 1〉.

• The kernel of the network is the rearrangement of a bitonicsequence into a sorted sequence.

– Typeset by FoilTEX – 10

Page 12: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Sorting Networks: Bitonic Sort

• Let s = 〈a0, a1, . . . , an−1〉 be a bitonic sequence such that a0 ≤a1 ≤ . . . ≤ an/2−1 and an/2 ≥ an/2+1 ≥ . . . ≥ an−1.

• Consider the following subsequences of s:

s1 = 〈mina0, an/2, mina1, an/2+1, . . . , minan/2−1, an−1〉s2 = 〈maxa0, an/2, maxa1, an/2+1, . . . , maxan/2−1, an−1〉

(1)

• Note that s1 and s2 are both bitonic and each element of s1 isless that every element in s2.

• We can apply the procedure recursively on s1 and s2 to get thesorted sequence.

– Typeset by FoilTEX – 11

Page 13: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Sorting Networks: Bitonic SortOriginalsequence 3 5 8 9 10 12 14 20 95 90 60 40 35 23 18 01st Split 3 5 8 9 10 12 14 0 95 90 60 40 35 23 18 202nd Split 3 5 8 0 10 12 14 9 35 23 18 20 95 90 60 403rd Split 3 0 8 5 10 9 14 12 18 20 35 23 60 40 95 904th Split 0 3 5 8 9 10 12 14 18 20 23 35 40 60 90 95

Merging a 16-element bitonic sequence through a series of log 16bitonic splits.

– Typeset by FoilTEX – 12

Page 14: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Sorting Networks: Bitonic Sort

• We can easily build a sorting network to implement this bitonicmerge algorithm.

• Such a network is called a bitonic merging network .

• The network contains log n columns. Each column contains n/2comparators and performs one step of the bitonic merge.

• We denote a bitonic merging network with n inputs by ⊕BM[n].

• Replacing the ⊕ comparators by comparators results in adecreasing output sequence; such a network is denoted byBM[n].

– Typeset by FoilTEX – 13

Page 15: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Sorting Networks: Bitonic Sort

18

23

35

40

60

90

95

20

14

12

10

9

8

5

3

95

90

60

40

35

23

20

18

14

12

10

9

8

5

3

0

90

95

40

60

23

35

20

18

12

14

9

10

5

8

0

3

40

90

60

95

20

18

23

35

9

0

12

10

0

5

8

33

5

10

14

8

9

12

14

0

95

90

40

35

23

20

18

60

0011

1100

1110

1101

1011

1010

1001

1000

0111

0110

0101

0100

0010

0001

0000

Wires

1111

A bitonic merging network for n = 16. The input wires arenumbered 0, 1 . . . , n − 1, and the binary representation of these

numbers is shown. Each column of comparators is drawnseparately; the entire figure represents a ⊕BM[16] bitonic

merging network. The network takes a bitonic sequence andoutputs it in sorted order.

– Typeset by FoilTEX – 14

Page 16: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Sorting Networks: Bitonic Sort

How do we sort an unsorted sequence using a bitonic merge?

• We must first build a single bitonic sequence from the givensequence.

• A sequence of length 2 is a bitonic sequence.

• A bitonic sequence of length 4 can be built by sorting the firsttwo elements using ⊕BM[2] and next two, using BM[2].

• This process can be repeated to generate larger bitonicsequences.

– Typeset by FoilTEX – 15

Page 17: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Sorting Networks: Bitonic Sort

BM[2]

BM[2]

BM[2]

BM[2]

BM[2]

BM[2]

BM[2]

BM[2]

BM[16]

BM[4]

BM[4]

BM[4]

BM[4]

BM[8]

BM[8]

0001

0100

0101

0000

0010

0011

0110

0111

1000

1001

10101011

1100

1101

1110

1111

Wires

A schematic representation of a network that converts an inputsequence into a bitonic sequence. In this example, ⊕BM[k] andBM[k] denote bitonic merging networks of input size k that use⊕ and comparators, respectively. The last merging network

(⊕BM[16]) sorts the input. In this example, n = 16.

– Typeset by FoilTEX – 16

Page 18: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Sorting Networks: Bitonic Sort

23

90

60

40

0

3

8

12

14

20

10

9

5

0

18

23

35

40

60

90

95

20

14

12

10

9

8

18

95

5

3

18

95

35

23

40

60

90

0

12

14

8

3

5

9

20

10

18

95

35

23

40

60

0

90

14

12

8

3

9

5

20

10

35

1000

1111

1110

1101

1100

1011

1010

1001

0111

0110

0101

0100

0011

0010

0001

0000

Wires

The comparator network that transforms an input sequence of16 unordered numbers into a bitonic sequence.

– Typeset by FoilTEX – 17

Page 19: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Sorting Networks: Bitonic Sort

• The depth of the network is Θ(log2 n).

• Each stage of the network contains n/2 comparators. Aserial implementation of the network would have complexityΘ(n log2 n).

– Typeset by FoilTEX – 18

Page 20: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Mapping Bitonic Sort to Hypercubes

• Consider the case of one item per processor. The questionbecomes one of how the wires in the bitonic network shouldbe mapped to the hypercube interconnect.

• Note from our earlier examples that the compare-exchangeoperation is performed between two wires only if their labelsdiffer in exactly one bit!

• This implies a direct mapping of wires to processors. Allcommunication is nearest neighbor!

– Typeset by FoilTEX – 19

Page 21: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Mapping Bitonic Sort to Hypercubes

Step 2

Step 4Step 3

Step 1

1101

1100

1010

1110

1111

1011

0000

0001

0101

0100 0110

0011

0111

1000

1001

0010

0100 1100

1101

1110

0000

0001

0101

0010

0110

0111

0011

1001

1000

1010

1011

1111

1100

1101

1010

0100

0000

0001

0101

0010

0110

0111

0011

1000

1001 1011

1111

1110

1100

1101

0100

0000

0001

0101

0010

0110

0111

1000

1001 1011

1111

1110

0011

1010

Communication during the last stage of bitonic sort. Each wire ismapped to a hypercube process; each connection represents a

compare-exchange between processes.

– Typeset by FoilTEX – 20

Page 22: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Mapping Bitonic Sort to Hypercubes

1110

11001101

1011

10011000

1010

01100101

0111

0100001100100001

1111

0000

Stage 3

1

1

1

1

1

1

1

1

2,1

2,1

2,1

2,1

3,2,1

3,2,1

4,3,2,1

Processors

Stage 4Stage 2Stage 1

Communication characteristics of bitonic sort on a hypercube.During each stage of the algorithm, processes communicate

along the dimensions shown.

– Typeset by FoilTEX – 21

Page 23: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Mapping Bitonic Sort to Hypercubes

1. procedure BITONIC SORT(label, d)

2. begin3. for i := 0 to d − 1 do4. for j := i downto 0 do5. if (i + 1)st bit of label 6= jth bit of label then6. comp exchange max(j);7. else8. comp exchange min(j);9. end BITONIC SORT

Parallel formulation of bitonic sort on a hypercube with n = 2d processes.

– Typeset by FoilTEX – 22

Page 24: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Mapping Bitonic Sort to Hypercubes

• During each step of the algorithm, every process performsa compare-exchange operation (single nearest neighborcommunication of one word).

• Since each step takes Theta(1) time, the parallel time is

TP = Θ(log2 n) (2)

• This algorithm is cost optimal w.r.t. its serial counterpart, but notw.r.t. the best sorting algorithm.

– Typeset by FoilTEX – 23

Page 25: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Mapping Bitonic Sort to Meshes

• The connectivity of a mesh is lower than that of a hypercube,so we must expect some overhead in this mapping.

• Consider the row-major shuffled mapping of wires toprocessors.

– Typeset by FoilTEX – 24

Page 26: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Mapping Bitonic Sort to Meshes

1101

1001 1010

0100

1000 1011

1100 1110 1111

0101 0110 0111

0000 0001 0010 0011

1110

1001

0010

1010

1101 1100

0000 0001 0011

0111 0110 0101 0100

1000 1011

1111

1000

1110

1100

1111

0000 0001 0100 0101

0010 0011 0110 0111

1001 1101

1010 1011

(a) (b) (c)

Different ways of mapping the input wires of the bitonic sortingnetwork to a mesh of processes: (a) row-major mapping, (b)

row-major snakelike mapping, and (c) row-major shuffledmapping.

– Typeset by FoilTEX – 25

Page 27: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Mapping Bitonic Sort to Meshes

Step 1 Step 2 Step 3 Step 4

Stage 4

The last stage of the bitonic sort algorithm for n = 16 on a mesh,using the row-major shuffled mapping. During each step, process

pairs compare-exchange their elements. Arrows indicate thepairs of processes that perform compare-exchange operations.

– Typeset by FoilTEX – 26

Page 28: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Mapping Bitonic Sort to Meshes

• In the row-major shuffled mapping, wires that differ at the ith

least-significant bit are mapped onto mesh processes that are2b(i−1)/2c communication links away.

• The total amount of communication performed by eachprocess is

∑log ni=1

∑ij=1 2b(j−1)/2c ≈ 7

√n, or Θ(

√n). The total

computation performed by each process is Θ(log2 n).

• The parallel runtime is:

TP =

comparisons︷ ︸︸ ︷

Θ(log2 n) +

communication︷ ︸︸ ︷

Θ(√

n).

• This is not cost optimal.

– Typeset by FoilTEX – 27

Page 29: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Block of Elements Per Processor

• Each process is assigned a block of n/p elements.

• The first step is a local sort of the local block.

• Each subsequent compare-exchange operation is replacedby a compare-split operation.

• We can effectively view the bitonic network as having (1 +log p)(log p)/2 steps.

– Typeset by FoilTEX – 28

Page 30: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Block of Elements Per Processor: Hypercube

• Initially the processes sort their n/p elements (using merge sort)in time Θ((n/p) log(n/p)) and then perform Θ(log2 p) compare-split steps.

• The parallel run time of this formulation is

TP =

local sort︷ ︸︸ ︷

Θ

(n

plog

n

p

)

+

comparisons︷ ︸︸ ︷

Θ

(n

plog2 p

)

+

communication︷ ︸︸ ︷

Θ

(n

plog2 p

)

.

• Comparing to an optimal sort, the algorithm can efficiently useup to p = Θ(2

√log n) processes.

• The isoefficiency function due to both communication andextra work is Θ(plog p log2 p).

– Typeset by FoilTEX – 29

Page 31: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Block of Elements Per Processor: Mesh

• The parallel runtime in this case is given by:

TP =

local sort︷ ︸︸ ︷

Θ

(n

plog

n

p

)

+

comparisons︷ ︸︸ ︷

Θ

(n

plog2 p

)

+

communication︷ ︸︸ ︷

Θ

(n√p

)

• This formulation can efficiently use up to p = Θ(log2 n) processes.

• The isoefficiency function is Θ(2√

p√p).

– Typeset by FoilTEX – 30

Page 32: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Performance of Parallel Bitonic Sort

The performance of parallel formulations of bitonic sort for nelements on p processes.

Maximum Number of Corresponding IsoefficiencyArchitecture Processes for E = Θ(1) Parallel Run Time Function

Hypercube Θ(2√

log n) Θ(n/(2√

log n) log n) Θ(plog p log2 p)

Mesh Θ(log2 n) Θ(n/ log n) Θ(2√

p√p)

Ring Θ(log n) Θ(n) Θ(2pp)

– Typeset by FoilTEX – 31

Page 33: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Bubble Sort and its Variants

The sequential bubble sort algorithm compares and exchangesadjacent elements in the sequence to be sorted:

1. procedure BUBBLE SORT(n)

2. begin3. for i := n − 1 downto 1 do4. for j := 1 to i do5. compare-exchange(aj, aj+1);6. end BUBBLE SORT

Sequential bubble sort algorithm.

– Typeset by FoilTEX – 32

Page 34: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Bubble Sort and its Variants

• The complexity of bubble sort is Θ(n2).

• Bubble sort is difficult to parallelize since the algorithm has noconcurrency.

• A simple variant, though, uncovers the concurrency.

– Typeset by FoilTEX – 33

Page 35: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Odd-Even Transposition

1. procedure ODD-EVEN(n)

2. begin3. for i := 1 to n do4. begin5. if i is odd then6. for j := 0 to n/2 − 1 do7. compare-exchange(a2j+1, a2j+2);8. if i is even then9. for j := 1 to n/2 − 1 do10. compare-exchange(a2j, a2j+1);11. end for12. end ODD-EVEN

Sequential odd-even transposition sort algorithm.

– Typeset by FoilTEX – 34

Page 36: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Odd-Even Transposition

3 3 4 5 6 8

Phase 6 (even)

2 1

53 2 8 6 4

2 3 8 5 6 4

5 48 62 3

3 5 8 4 6

3 5 4 8 6

3 4 5 6 8

Phase 1 (odd)

Unsorted

Phase 2 (even)

Phase 3 (odd)

Phase 4 (even)

Phase 5 (odd)

3 1

13

3 1

12

2

2

3

3

3

1

1

Sorted

3 3 4 5 6 8

3 3 4 5 6 8

1 2

1 2

Phase 8 (even)

Phase 7 (odd)

Sorting n = 8 elements, using the odd-even transposition sortalgorithm. During each phase, n = 8 elements are compared.

– Typeset by FoilTEX – 35

Page 37: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Odd-Even Transposition

• After n phases of odd-even exchanges, the sequence is sorted.

• Each phase of the algorithm (either odd or even) requires Θ(n)comparisons.

• Serial complexity is Θ(n2).

– Typeset by FoilTEX – 36

Page 38: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Parallel Odd-Even Transposition

• Consider the one item per processor case.

• There are n iterations, in each iteration, each processor doesone compare-exchange.

• The parallel run time of this formulation is Θ(n).

• This is cost optimal with respect to the base serial algorithm butnot the optimal one.

– Typeset by FoilTEX – 37

Page 39: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Parallel Odd-Even Transposition

1. procedure ODD-EVEN PAR(n)

2. begin3. id := process’s label4. for i := 1 to n do5. begin6. if i is odd then7. if id is odd then8. compare-exchange min(id + 1);9. else10. compare-exchange max(id − 1);11. if i is even then12. if id is even then13. compare-exchange min(id + 1);14. else15. compare-exchange max(id − 1);16. end for17. end ODD-EVEN PAR

Parallel formulation of odd-even transposition.

– Typeset by FoilTEX – 38

Page 40: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Parallel Odd-Even Transposition

• Consider a block of n/p elements per processor.

• The first step is a local sort.

• In each subsequent step, the compare exchange operation isreplaced by the compare split operation.

• The parallel run time of the formulation is

TP =

local sort︷ ︸︸ ︷

Θ

(n

plog

n

p

)

+

comparisons︷ ︸︸ ︷

Θ(n) +

communication︷ ︸︸ ︷

Θ(n).

– Typeset by FoilTEX – 39

Page 41: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Parallel Odd-Even Transposition

• The parallel formulation is cost-optimal for p = O(log n).

• The isoefficiency function of this parallel formulation is Θ(p 2p).

– Typeset by FoilTEX – 40

Page 42: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Shellsort

• Let n be the number of elements to be sorted and p be thenumber of processes.

• During the first phase, processes that are far away from eachother in the array compare-split their elements.

• During the second phase, the algorithm switches to an odd-even transposition sort.

– Typeset by FoilTEX – 41

Page 43: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Parallel Shellsort

• Initially, each process sorts its block of n/p elements internally.

• Each process is now paired with its corresponding process inthe reverse order of the array. That is, process Pi, where i < p/2,is paired with process Pp−i−1.

• A compare-split operation is performed.

• The processes are split into two groups of size p/2 each and theprocess repeated in each group.

– Typeset by FoilTEX – 42

Page 44: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Parallel Shellsort

0 123 4 5 6 7

0 123 4 5 6 7

0 123 4 5 6 7

An example of the first phase of parallel shellsort on aneight-process array.

– Typeset by FoilTEX – 43

Page 45: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Parallel Shellsort

• Each process performs d = log p compare-split operations.

• With O(p) bisection width, the each communication can beperformed in time Θ(n/p) for a total time of Θ((n log p)/p).

• In the second phase, l odd and even phases are performed,each requiring time Θ(n/p).

• The parallel run time of the algorithm is:

TP =

local sort︷ ︸︸ ︷

Θ

(n

plog

n

p

)

+

first phase︷ ︸︸ ︷

Θ

(n

plog p

)

+

second phase︷ ︸︸ ︷

Θ

(

ln

p

)

. (3)

– Typeset by FoilTEX – 44

Page 46: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Quicksort

• Quicksort is one of the most common sorting algorithms forsequential computers because of its simplicity, low overhead,and optimal average complexity.

• Quicksort selects one of the entries in the sequence to be thepivot and divides the sequence into two – one with all elementsless than the pivot and other greater.

• The process is recursively applied to each of the sublists.

– Typeset by FoilTEX – 45

Page 47: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Quicksort

1. procedure QUICKSORT (A, q, r)

2. begin3. if q < r then4. begin5. x := A[q];6. s := q;7. for i := q + 1 to r do8. if A[i] ≤ x then9. begin10. s := s + 1;11. swap(A[s], A[i]);12. end if13. swap(A[q], A[s]);14. QUICKSORT (A, q, s);15. QUICKSORT (A, s + 1, r);16. end if17. end QUICKSORT

The sequential quicksort algorithm.

– Typeset by FoilTEX – 46

Page 48: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Quicksort

1 2 3 3 4 5 8 7

1 2 3 3 4 5 7 8

3 2 1 5 8 4 3 7(a)

(b)

(c)

(d)

(e)

1 2 3 5 8 4 3 7

1 2 3 3 4 5 7 8

Final position

Pivot

Example of the quicksort algorithm sorting a sequence of sizen = 8.

– Typeset by FoilTEX – 47

Page 49: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Quicksort

• The performance of quicksort depends critically on the qualityof the pivot.

• In the best case, the pivot divides the list in such a way that thelarger of the two lists does not have more than αn elements (forsome constant α).

• In this case, the complexity of quicksort is O(n log n).

– Typeset by FoilTEX – 48

Page 50: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Parallelizing Quicksort

• Lets start with recursive desomposition – the list is partitionedserially and each of the subproblems is handled by a differentprocessor.

• The time for this algorithm is lower-bounded by Ω(n)!

• Can we parallelize the partitioning step – in particular, if we canuse n processors to partition a list of length n around a pivot inO(1) time, we have a winner.

• This is difficult to do on real machines, though.

– Typeset by FoilTEX – 49

Page 51: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Parallelizing Quicksort: PRAM Formulation

• We assume a CRCW (concurrent read, concurrent write)PRAM with concurrent writes resulting in an arbitrary writesucceeding.

• The formulation works by creating pools of processors. Everyprocessor is assigned to the same pool initially and has oneelement.

• Each processor attempts to write its element to a commonlocation (for the pool).

• Each processor tries to read back the location. If the valueread back is greater than the processor’s value, it assigns itselfto the ‘left’ pool, else, it assigns itself to the ‘right’ pool.

• Each pool performs this operation recursively.

• Note that the algorithm generates a tree of pivots. The depthof the tree is the expected parallel runtime. The average valueis O(log n).

– Typeset by FoilTEX – 50

Page 52: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Parallelizing Quicksort: PRAM Formulation3

5

3

7

1

2

4

8

A binary tree generated by the execution of the quicksortalgorithm. Each level of the tree represents a different

array-partitioning iteration. If pivot selection is optimal, then theheight of the tree is Θ(log n), which is also the number of

iterations.

– Typeset by FoilTEX – 51

Page 53: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Parallelizing Quicksort: PRAM Formulation1 2 3 4 5 6 7 8

1 2 3 4 5 6 7 8 1 2 3 4 5 6 7 8

26

73

5 8

2 3 6 7

3 7

8

1 2 3 4 5 6 7 8

[4] 54

[1] 33

[6] 33

[5] 82

[2] 21

[3] 13 [7] 40

[8] 72

54 82 40 72132133 33(a)

(b)

(c)

1

5

(f)

(d) (e)

2

6

1

5

8 2

6

3 1

5

8

7

leftchild

rightchild

leftchild

rightchild

leftchild

rightchild

1

root = 4

The execution of the PRAM algorithm on the array shown in (a).

– Typeset by FoilTEX – 52

Page 54: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Parallelizing Quicksort: Shared Address SpaceFormulation

• Consider a list of size n equally divided across p processors.

• A pivot is selected by one of the processors and made knownto all processors.

• Each processor partitions its list into two, say Li and Ui, basedon the selected pivot.

• All of the Li lists are merged and all of the Ui lists are mergedseparately.

• The set of processors is partitioned into two (in proportion of thesize of lists L and U). The process is recursively applied to eachof the lists.

– Typeset by FoilTEX – 53

Page 55: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Shared Address Space Formulation

after globalrearrangement

pivot=7

pivot selection

after localrearrangement

Firs

t Ste

p

after localrearrangement

Four

th S

tep

pivot=5 pivot=17

after localrearrangement

after globalrearrangement

after localrearrangement

after globalrearrangement

Seco

nd S

tep

pivot selection

Thi

rd S

tep

pivot selection

pivot=11

Solution

7 2 1 6 3 4 5 18 13 17 14 20 10 15 9 19 16 12 11 8

1 313 1014 920177 418 152 196 1116 12 5 8

314 9207 15 1962 18 13 1 17 10 4 16 5 1112 8

10 9 8 12 11 13 17 1514 16

7 2 1 6 3 4 5 18 13 17 14 20 10 15 9 19 16 12 11 8

2 6 3 4 571 13 20 10 15 9 19 16 12 11 81714 18

21 13 17143 4 5 67 10 15 9 16 12 11 8 18 20 19

10 9 8 12 11 13 17 1514 16

21 13 17143 4 5 67 10 15 9 16 12 11 8 18 20 19

21 3 4 5 6 7 18 19 2013 17 15 9 12 111410 8 16

21 3 4 5 6 7 18 19 2013 14 15 16 179 108 11 12

PSfrag replacements

P0

P0

P0

P0

P0

P0

P0

P1

P1

P1

P1

P1

P1

P1

P2

P2

P2

P2

P2

P2

P2

P2

P3

P3

P3

P3

P3

P3

P3

P3

P4

P4

P4

P4

P4

P4

P4

– Typeset by FoilTEX – 54

Page 56: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Parallelizing Quicksort: Shared Address SpaceFormulation

• The only thing we have not described is the globalreorganization (merging) of local lists to form L and U .

• The problem is one of determining the right location for eachelement in the merged list.

• Each processor computes the number of elements locally lessthan and greater than pivot.

• It computes two sum-scans to determine the starting locationfor its elements in the merged L and U lists.

• Once it knows the starting locations, it can write its elementssafely.

– Typeset by FoilTEX – 55

Page 57: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Parallelizing Quicksort: Shared Address SpaceFormulation

pivot=7

pivot selection

after localrearrangement

after globalrearrangement

1 313 1014 920177 418 152 196 1116 12 5 8

314 9207 15 1962 18 13 1 17 10 4 16 5 1112 8

2 1 6 3 4 5 18 13 17 14 20 10 15 9 19 16 12 11 870 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19

2 1 1 2 1 2 3 3 2 3

0 2 5 8 100 2 3 4 6 7 13

PSfrag replacements

P0

P0

P1

P1

P2

P2

P3

P3

P4

P4

|Si| |Li|

PrefixSum of |Si|PrefixSum of |Li|

Prefix SumPrefix Sum

Efficient global rearrangement of the array.

– Typeset by FoilTEX – 56

Page 58: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Parallelizing Quicksort: Shared Address SpaceFormulation

• The parallel time depends on the split and merge time, and thequality of the pivot.

• The latter is an issue independent of parallelism, so we focus onthe first aspect, assuming ideal pivot selection.

• The algorithm executes in four steps: (i) determine andbroadcast the pivot; (ii) locally rearrange the array assignedto each process; (iii) determine the locations in the globallyrearranged array that the local elements will go to; and (iv)perform the global rearrangement.

• The first step takes time Θ(log p), the second, Θ(n/p), the third,Θ(log p), and the fourth, Θ(n/p).

• The overall complexity of splitting an n-element array is Θ(n/p)+Θ(log p).

– Typeset by FoilTEX – 57

Page 59: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Parallelizing Quicksort: Shared Address SpaceFormulation

• The process recurses until there are p lists, at which point, thelists are sorted locally.

• Therefore, the total parallel time is:

TP =

local sort︷ ︸︸ ︷

Θ

(n

plog

n

p

)

+

array splits︷ ︸︸ ︷

Θ

(n

plog p

)

+ Θ(log2 p). (4)

• The corresponding isoefficiency is Θ(p log2 p) due to broadcastand scan operations.

– Typeset by FoilTEX – 58

Page 60: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Parallelizing Quicksort: Message Passing Formulation

• A simple message passing formulation is based on the recursivehalving of the machine.

• Assume that each processor in the lower half of a p processorensemble is paired with a corresponding processor in the upperhalf.

• A designated processor selects and broadcasts the pivot.

• Each processor splits its local list into two lists, one less (Li), andother greater (Ui) than the pivot.

• A processor in the low half of the machine sends its list Ui to thepaired processor in the other half. The paired processor sendsits list Li.

• It is easy to see that after this step, all elements less than thepivot are in the low half of the machine and all elementsgreater than the pivot are in the high half.

– Typeset by FoilTEX – 59

Page 61: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Parallelizing Quicksort: Message Passing Formulation

• The above process is recursed until each processor has its ownlocal list, which is sorted locally.

• The time for a single reorganization is Θ(log p) for broadcastingthe pivot element, Θ(n/p) for splitting the locally assignedportion of the array, Θ(n/p) for exchange and localreorganization.

• We note that this time is identical to that of the correspondingshared address space formulation.

• It is important to remember that the reorganization of elementsis a bandwidth sensitive operation.

– Typeset by FoilTEX – 60

Page 62: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Bucket and Sample Sort

• In Bucket sort, the range [a, b] of input numbers is divided into mequal sized intervals, called buckets.

• Each element is placed in its appropriate bucket.

• If the numbers are uniformly divided in the range, the bucketscan be expected to have roughly identical number ofelements.

• Elements in the buckets are locally sorted.

• The run time of this algorithm is Θ(n log(n/m)).

– Typeset by FoilTEX – 61

Page 63: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Parallel Bucket Sort

• Parallelizing bucket sort is relatively simple. We can select m =p.

• In this case, each processor has a range of values it isresponsible for.

• Each processor runs through its local list and assigns each of itselements to the appropriate processor.

• The elements are sent to the destination processors using asingle all-to-all personalized communication.

• Each processor sorts all the elements it receives.

– Typeset by FoilTEX – 62

Page 64: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Parallel Bucket and Sample Sort

• The critical aspect of the above algorithm is one of assigningranges to processors. This is done by suitable splitter selection.

• The splitter selection method divides the n elements into mblocks of size n/m each, and sorts each block by usingquicksort.

• From each sorted block it chooses m − 1 evenly spacedelements.

• The m(m − 1) elements selected from all the blocks representthe sample used to determine the buckets.

• This scheme guarantees that the number of elements endingup in each bucket is less than 2n/m.

– Typeset by FoilTEX – 63

Page 65: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Parallel Bucket and Sample SortInitial element distribution

Local sort &sample selection

Global splitter selection

Final elementassignment

Sample combining

113 1014 201718 2 6722 24 3 191615 23 4 11 12 5 8219

1 2 13 181714 22 3 6 9 2410 15 20 21 4 5 8 11 12 16 23197

7 17 9 20 8 16

7 8 9 16 17 20

1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 20 21 22 23 2419

PSfrag replacements

P0

P0

P0

P1

P1

P1

P2

P2

P2

An example of the execution of sample sort on an array with 24elements on three processes.

– Typeset by FoilTEX – 64

Page 66: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Parallel Bucket and Sample Sort

• The splitter selection scheme can itself be parallelized.

• Each processor generates the p − 1 local splitters in parallel.

• All processors share their splitters using a single all-to-allbroadcast operation.

• Each processor sorts the p(p−1) elements it receives and selectsp − 1 uniformly spaces splitters from them.

– Typeset by FoilTEX – 65

Page 67: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Parallel Bucket and Sample Sort: Analysis

• The internal sort of n/p elements requires time Θ((n/p) log(n/p)),and the selection of p − 1 sample elements requires time Θ(p).

• The time for an all-to-all broadcast is Θ(p2), the time to internallysort the p(p−1) sample elements is Θ(p2 log p), and selecting p−1evenly spaced splitters takes time Θ(p).

• Each process can insert these p − 1 splitters in its local sortedblock of size n/p by performing p − 1 binary searches in timeΘ(p log(n/p)).

• The time for reorganization of the elements is O(n/p).

– Typeset by FoilTEX – 66

Page 68: Sorting Algorithmskarypis/parbook/Lectures/AG/chap9_slides.pdf · Sorting Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text fiIntroduction

Parallel Bucket and Sample Sort: Analysis

• The total time is given by:

TP =

local sort︷ ︸︸ ︷

Θ

(n

plog

n

p

)

+

sort sample︷ ︸︸ ︷

Θ(p2 log p

)+

block partition︷ ︸︸ ︷

Θ

(

p logn

p

)

+

communication︷ ︸︸ ︷

Θ(n/p). (5)

• The isoefficiency of the formulation is Θ(p3 log p).

– Typeset by FoilTEX – 67