28
Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia [email protected]

Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia

Embed Size (px)

DESCRIPTION

UVA work items - 3 tracks Provisioning across CHEETAH and UltraScience networks Transport protocol for dedicated circuits Extend CHEETAH concept to enable heterogeneous connections – “connection-oriented internet”

Citation preview

Page 1: Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia

Enabling Supernova Computations on Dedicated Channels

Malathi VeeraraghavanUniversity of [email protected]

Page 2: Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia

Acknowledgment• Thanks to the DOE MICS R&D program

– DOE DE-FG02-04ER25640• Graduate students

– Zhanxiang Huang– Anant P. Mudambi– Xiuduan Fang– Xiangfei Zhu

• Postdoc fellow– Xuan Zheng

• CHEETAH project co-PIs: – ORNL, NCSU CUNY

Page 3: Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia

UVA work items - 3 tracks• Provisioning across CHEETAH and

UltraScience networks• Transport protocol for dedicated

circuits• Extend CHEETAH concept to enable

heterogeneous connections – “connection-oriented internet”

Page 4: Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia

CHEETAH Network(data-plane)

Cisco7600 switch

1GCompute-0-4

Orbitty Compute Nodes

1GOC192 OC192 GbE

1-8-331-8-341-8-351-8-36

1-6-1

1-6-171-8-37

Cisco7600 switch

HHHH

H 1G1G1G1G

1-7-1

Compute-0-3Compute-0-2Compute-0-1Compute-0-0

1G1G1G1G

wukong H

1G1-8-381-7-17

cheetah-nc

5x1Gbps VLAN

OC192

1-6-1

1-6-17

10GbE

1-7-1

GbE1-7-331-7-341-7-351-7-361-7-371-7-381-7-39

1GZelda1

HH

H1G1GZelda2

Zelda31G

1G

Zelda4HH

Zelda5

2x1Gbps MPLS tunnels

1G1G

Cheetah-atl

OC-192 lamda

10GbEGbE1-7-331-7-341-7-351-7-36

Cheetah-ornl

1-7-1 1-6-1

OC192

X1(E)UCNS1GFC1G

1G

Juniper320

router

Juniper320

router

1G

1G

Force10E300

switch

ORNL

Atlanta

NC

Direct fibers

VLANsMPLS tunnels

1Gbps VLAN

USN

Page 5: Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia

CHEETAH nodes at the three PoPs• Sycamore SN16000 intelligent

optical switch– GbE, 10GbE, and SONET interface

cards (we have OC192s)– Switch OS - BroadLeaf – implements

GMPLS protocols since Release 7.0• Pure SONET circuits – excellent

GMPLS signaling and routing support• Ethernet-to-SONET GMPLS standards

not officially released yet• But Sycamore provided us a

proprietary solution - it is quite stable

Page 6: Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia

Distributed signaling implemented(RSVP client software for end host done)

SN16000

GbE ControlCard OC192

SN16000

GbE ControlCard OC192

EndHost

NIC1NIC2 End

Host

NIC1NIC2

RSVP-TE Path

RSVP-TE PathRSVP-TE Path

BroadLeaf MessageRSVP-TE Resv

bwlib integrated into FTP codefor automatic initiation of circuitby end application software

Page 7: Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia

Call setup delay measurementsCircuittype

End-to-endcircuit setupdelay (s)

PATHprocessingdelay atthe SN16k-NC

RESVprocessingdelay atthe SN16k-NC

OC1 0.166103 0.091119 0.008689

OC3 0.165450 0.090852 0.008650

1Gbps Ethernet/EoS

4.982071 4.907113 0.008697

21 OC1 on SONET side are set up one at a time21 x 0.166 + 21 x 0.025 = ~4sec

propagation + emission delays

Page 8: Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia

Key results• Distributed GMPLS signaling + routing

using off-the-shelf switches works well!• GMPLS control-plane protocols suitable for

– Only call blocking mode of bandwidth sharing– Immediate-request calls, not book-ahead

• If the number of circuits (m) sharing the link is small (order of 10), e.g. 1Gbps on 10Gbps link– Either call blocking probability will be high– Or link utilization will be low

Page 9: Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia

Call blocking probability as a function of utilization, U, and number of circuits, m

Page 10: Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia

UVA work items - 3 tracks• Provisioning across CHEETAH and

UltraScience networksTransport protocol for dedicated

circuits• Extend CHEETAH concept to enable

heterogeneous connections – “connection-oriented internet”

Page 11: Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia

FTP over TCP across circuitseems to do best!

zelda4 - compute-0-0

zelda4 - zelda3

zelda4 - wukong

zelda3 - compute-0-0

zelda3 - wukong

wukong - compute-0-0

RTT (ms) 32 13.7 8.75 1

Memory-to-memory transfers

Iperf TCP 938 924 938 938 931 900 933 934 934 N/A N/A 933

Iperf UDP 888 913 957 957 646 830 800 913 653 727 750 645

Disk-to-disk transfers (1.3GB file)

FTP 752 552 585 585 702 458 878 479 702 722* N/A 620

SFTP 25 18.8 34.9 35.1 17.3 17.9 41 18.7 26.3 24.3 26.4 130

SABUL 640 770 488 624 470 404 638 848 463 610 520 479

Hurricane 524 545 537 422 456 368 530 542 264 282 N/A N/A

BBCP N/A 500 607 657 N/A N/A N/A 611 425 379 87 513

FRTPv1 629 600 853 610 510 664 644 787 515 620 368 388

Page 12: Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia

First attempt• Fixed-Rate Transport Protocol (FRTP)

– User-space implementation– UDP-based solution– Null flow control

• Thought we could estimate receive rate (disk bottleneck)• Use this as circuit rate• Stream data from sender at this (fixed) circuit rate

– Hard to do the above because of• Variability in disk receive rate• Multitasking hosts (general-purpose)

– CPU-intensive solution for maintaining constant sending rate

Page 13: Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia

Decided to modify TCP• Reasons

– Due to failure of null flow control solution, we decided on window based flow control

• best out of three: ON/OFF and rate based• need kernel-level implementation for window control

– Self-clocking in TCP works well to maintain fixed sending rate

• Busy wait discarded due to high CPU utilization• Signals and timers unreliable

– TCP implementation + Web100 stable and fast

Page 14: Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia

Protocol Design• Congestion control: Control plane function

– TCP's data-plane congestion control redundant• Flow control: Window based• Error control: Required for reliable transfer

– Thought we could use just negative ACKs because of in-sequence delivery characteristic of circuits

– But cost of retrieving errored frames from disk high– Need positive ACKs to remove packets held in

retransmission buffers • Multiplexing: TCP's port solution usable with circuits• Conclusion: TCP’s mechanisms suitable for all but

congestion control

Page 15: Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia

Hence Circuit-TCPHow does it differ from TCP?

• Data plane– Slow Start/Congestion Avoidance removed– Sender sends at the fixed rate of circuit

• Control plane– TCP has three-way handshake (host-to-host)

to synchronize initial sequence numbers– Circuit-TCP determines rate to use for the

transfer and initiates request for circuit

Page 16: Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia

Implementation Issues• Selecting the rate of the circuit

– End-hosts usually a bottleneck, esp., for disk-to-disk transfers. Difficult to pin down a constant bottleneck rate for the whole transfer

– Pragmatic approach: • Minimize sources of variability (e.g., avoid

multitasking)• Use an empirical estimate of the receiver’s disk

write rate• Maintaining the sending rate at the circuit

rate

Page 17: Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia

Linux implementationSender:• Try to maintain fixed amount of outstanding data

in network

• In TCP implementation replace min(cwnd, rwnd) by min(ncap, rwnd)

• 2 requirements– C-TCP and TCP co-exist– If socket is C-TCP then the kernel needs to know ncap

for the socket• Web100 provides API that was extended to meet

these 2 requirements

ncap (>= BDP = circuit rate * RTT)

Page 18: Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia

Linux implementationReceiver:• Linux increments advertised window

in a slow start-like fashion• So for the initial few RTTs, at the

sender min(ncap, rwnd) = rwnd, leading to low circuit utilization

• To avoid this, for C-TCP, we modified this code

Page 19: Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia

Three sets of results• "Small" memory-to-memory transfers

– up to 100MB– 1Gbps circuit, RTT: 13.6ms

• How RTT varies - sustained sending rate– 500Mbps SONET circuit; RTT: 13.6ms– Sycamore gateway will buffer Ethernet

frames since sending NIC will send at 1Gbps• Disk-to-disk: 1.6GB over a 1Gbps circuit

Page 20: Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia

"Small" memory-to-memory transfersA

vera

ge T

hrou

ghpu

t (M

bps)

Amount of data transferred (KB)

Rel

ativ

e de

lay

Relative delayTCP

C-TCP

Page 21: Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia

RTT variation: Reno TCP

Time (seconds)

RTT

(m

s)Th

roug

hput

(M

bps)

Page 22: Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia

RTT variation: BIC-TCP

Time (seconds)

RTT

(m

s)Th

roug

hput

(M

bps)

Page 23: Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia

RTT variation: C-TCP

Time (seconds)

RTT

(m

s)Th

roug

hput

(M

bps)

Page 24: Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia

Disk-to-disk: 1.6GB; 1Gbps circuit

Experiment run number

Thro

ughp

ut (M

bps)

TCPC-TCP

Page 25: Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia

UVA work items• Provisioning across CHEETAH and

UltraScience networks• Transport protocol for dedicated

circuits: Fixed-Rate Transport Protocol (FRTP)

Extend CHEETAH concept to enable heterogeneous connections – “connection-oriented internet”

Page 26: Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia

Connection-oriented internet• Cisco and Juniper routers implement

– MPLS (user-plane)– RSVP-TE (control-plane)

• Many vendors’ Ethernet switches– Untagged (port-mapped) VLANs– Tagged VLANs with priority queueing– Add external GMPLS controller

Page 27: Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia

Design of signaling procedures• Key idea

– Signaling message encapsulation• Decreases delay - multiple RTTs avoided

– Combine with Interface Adaptation Capability Descriptor (IACD)

• CSPF not very useful– TE-LSAs only intra-area within OSPF– Inter-domain - topology hiding

Page 28: Enabling Supernova Computations on Dedicated Channels Malathi Veeraraghavan University of Virginia

Recently completed papers• A. P. Mudambi, X. Zheng, and M. Veeraraghavan, “A

transport protocol for dedicated circuits, submitted to IEEE ICC 2006.

• X. Zhu, X. Zheng, M. Veeraraghavan, Z. Li, Q. Song, I. Habib N. S. V. Rao, “Implementation of a GMPLS-based Network with End Host Initiated Signaling,” submitted to IEEE ICC 2006.

• M. Veeraraghavan, X. Fang, X. Zheng, “On the suitability of applications for GMPLS networks,” submitted to IEEE ICC 2006.

• M. Veeraraghavan, X. Zheng, Z. Huang, “On the use of GMPLS networks to support Grid Computing,” submitted to IEEE Communications Magazine.