41
Kirsten Fagnan NERSC User Services Februar 12, 2013 Getting Started at NERSC

Kirsten Fagnan NERSC User Services Februar 12, 2013

  • Upload
    shlomo

  • View
    42

  • Download
    2

Embed Size (px)

DESCRIPTION

Getting Started at NERSC. Kirsten Fagnan NERSC User Services Februar 12, 2013. Purpose. This presentation will help you get familiar with NERSC and its facilities Practical information Introduction to terms and acronyms This is not a programming tutorial - PowerPoint PPT Presentation

Citation preview

Page 1: Kirsten Fagnan NERSC User Services Februar  12,  2013

Kirsten FagnanNERSC User ServicesFebruar 12, 2013

Getting Started at NERSC

Page 2: Kirsten Fagnan NERSC User Services Februar  12,  2013

Purpose• This presentation will help you get familiar with

NERSC and its facilities– Practical information– Introduction to terms and acronyms

• This is not a programming tutorial– But you will learn how to get help and what kind of help is

available– We can give presentations on programming languages and

parallel libraries – just ask

Page 3: Kirsten Fagnan NERSC User Services Februar  12,  2013

Outline• What is NERSC?• Computing Resources• Storage Resources• How to Get Help

3

Page 4: Kirsten Fagnan NERSC User Services Februar  12,  2013

What is NERSC?

4

Page 5: Kirsten Fagnan NERSC User Services Februar  12,  2013

NERSC Leads DOE in Scientific Computing Productivity

5

NERSC computing for science• 5500 users, 600 projects• From 48 states; 65% from universities • Hundreds of users each day• 1,500 publications per yearSystems designed for science• 1.3PF Petaflop Cray system, Hopper

- Among Top 20 fastest in world- Fastest open Cray XE6 system- Additional clusters, data storage

Page 6: Kirsten Fagnan NERSC User Services Februar  12,  2013

We support a diverse workload

- 6 -

• Many codes (600+) and algorithms

• Computing at Scale and at High Volume

2012 Job Size Breakdown on Hopper

Jan 2012

Dec 2012

Frac

tion

of

hour

s

1

0.8

0.6

0.4

0.2

0

Page 7: Kirsten Fagnan NERSC User Services Februar  12,  2013

Our operational priority is providing highly available HPC resources backed by exceptional user support

• We maintain a very high availability of resources (>90%)– One large HPC system is available

at all times to run large-scale simulations and solve high throughput problems

• Our goal is to maximize the productivity of our users– One-on-one consulting– Training (e.g., webinars)– Extensive use of web pages– We solve or have a path to solve

80% of user tickets within 3 business days

- 7 -

Page 8: Kirsten Fagnan NERSC User Services Februar  12,  2013

Computing Resources

8

Page 9: Kirsten Fagnan NERSC User Services Februar  12,  2013

Current NERSC Systems

9

Large-Scale Computing SystemsHopper (NERSC-6): Cray XE6 • 6,384 compute nodes, 153,216 cores• 144 Tflop/s on applications; 1.3 Pflop/s peakEdison (NERSC-7): Cray Cascade• To be operational in 2013• Over 200 Tflop/s on applications, 2 Pflop/s peak

HPSS Archival Storage• 240 PB capacity• 5 Tape libraries• 200 TB disk cache

NERSC Global Filesystem (NGF)Uses IBM’s GPFS

• 8.5 PB capacity• 15GB/s of bandwidth

Midrange 140 Tflops totalCarver

• IBM iDataplex cluster• 9884 cores; 106TF

PDSF (HEP/NP)• ~1K core cluster

GenePool (JGI)• ~5K core cluster• 2.1 PB Isilon File System

Analytics & Testbeds

Dirac 48 Fermi GPU nodes

Page 10: Kirsten Fagnan NERSC User Services Februar  12,  2013

Structure of the Genepool System

- 10 -

compute nodes

gpint nodes

high priority

& interacti

ve nodesfpga

web services

database services

login nodes

filesystems

ssh genepool.nersc.gov http://…jgi-psf.orgUser Access Command Line Scheduler Service

Page 11: Kirsten Fagnan NERSC User Services Februar  12,  2013

Compute Node Hardware

- 11 -

Count Cores Slots ScheduleableMemory

Memory/Slot

Interconnects

515 8 8 42G 5.25G 1Gb Ethernet

220 16 16 120G 7.5G 14x FDR Infiniband

8 24 24 252G 10.5G 1G Ethernet

9 32 32 500G 15.625G 4 have 10G Eth5 have Infiniband

3 32 32 1000G 31.25G 1 has 10G Eth2 have Infiniband

1 80 64 2000G 31.25G 10G Ethernet

• The 42G nodes are scheduled “by slot”• Multiple jobs can run on the same node at the same

time• Higher memory nodes are exclusively scheduled

• Only one job can run at a time on a node

Page 12: Kirsten Fagnan NERSC User Services Februar  12,  2013

Hopper - Cray XE6

• 1.2 GB memory / core (2.5 GB / core on "fat" nodes) for applications

• /scratch disk quota of 5 TB• 2 PB of /scratch disk• Choice of full Linux operating

system or optimized Linux OS (Cray Linux)

• PGI, Cray, Pathscale, GNU compilers

12

• 153,408 cores, 6,392 nodes• "Gemini” interconnect• 2 12-core AMD 'MagnyCours'

2.1 GHz processors per node• 24 processor cores per node• 32 GB of memory per node

(384 "fat" nodes with 64 GB)• 216 TB of aggregate memory

Use Hopper for your biggest, most computationally challenging problems.

Page 13: Kirsten Fagnan NERSC User Services Februar  12,  2013

Edison - Cray XC30 (Phase 1 / 2)

• 4 / TBA GB memory / core for applications

• 1.6 / 6.4 PB of /scratch disk• CCM compatibility mode available• Intel, Cray, GNU compilers

13

• 10K / 105K compute cores• ”Aires” interconnect• 2 8-core Intel ’Sandy Bridge' 2.6

GHz processors per node (Phase 2 TBA)

• 16 / TBA processor cores per node• 64 GB of memory per node• 42 / 333 TB of aggregate memory• Edison Phase 1 access Feb. or

March 2013 / Phase 2 in Summer or Fall 2013

Use Edison for your most computationally challenging problems.

Page 14: Kirsten Fagnan NERSC User Services Februar  12,  2013

Carver - IBM iDataPlex • 3,200 compute cores• 400 compute nodes• 2 quad-core Intel Nehalem 2.67 GHz

processors per node• 8 processor cores per node• 24 GB of memory per node (48 GB on

80 "fat" nodes)• 2.5 GB / core for applications (5.5

GB / core on "fat" nodes)• InfiniBand 4X QDR

14

• NERSC global /scratch directory quota of 20 TB

• Full Linux operating system• PGI, GNU, Intel compilers

Use Carver for jobs that use up to 512 cores, need a fast CPU, need a standard Linux configuration, or need up to 48 GB of memory on a node.

Page 15: Kirsten Fagnan NERSC User Services Februar  12,  2013

Dirac – GPU Computing Testbed• 50 GPUs• 50 compute nodes

• 2 quad-core Intel Nehalem 2.67 GHz processors

• 24 GB DRAM memory

• 44 nodes: 1 NVIDIA Tesla C2050 (Fermi) GPU with 3GB of memory and 448 cores

• 1 node: 4 NVIDIA Tesla C2050 (Fermi) GPU's, each with 3GB of memory and 448 processor cores.

• InfiniBand 4X QDR

15

• CUDA 5.0, OpenCL, PGI and HMPP directives

• DDT CUDA-enabled debugger• PGI, GNU, Intel compilers

Use Dirac for developing and testing GPU codes.

Page 16: Kirsten Fagnan NERSC User Services Februar  12,  2013

How to Get Help

16

Page 17: Kirsten Fagnan NERSC User Services Februar  12,  2013

NERSC Services•NERSC’s emphasis is on enabling scientific discovery•User-oriented systems and services

–We think this is what sets NERSC apart from other centers

•Help Desk / Consulting–Immediate direct access to consulting staff that includes 7 Ph.Ds

•User group (NUG) has tremendous influence–Monthly teleconferences & yearly meetings

•Requirement-gathering workshops with top scientists –One each for the six DOE Program Offices in the Office of Science

–http://www.nersc.gov/science/requirements-workshops/

•Ask, and we’ll do whatever we can to fulfill your request

Page 18: Kirsten Fagnan NERSC User Services Februar  12,  2013

Your JGI Consultants

18

Seung-Jin Sul, Ph.D. Bioinformatics

Doug Jacobsen, Ph.D. Bioinformatics

Kirsten Fagnan, Ph.D. Applied Math

One of us is onsite every day! Stop by 400-413 if you have questions!

Page 19: Kirsten Fagnan NERSC User Services Februar  12,  2013

How to Get Helphttp://www.nersc.gov/

For Users sectionMy NERSC page (link at upper right)

1-800-666-3772 (or 1-510-486-8600)

Computer Operations* = menu option 1 (24/7)

Account Support (passwords) = menu option 2, [email protected]

HPC Consulting = menu option 3, or [email protected](8-5, M-F Pacific time)

Online Help Desk = https://help.nersc.gov/

* Passwords during non-business hours

19

Page 20: Kirsten Fagnan NERSC User Services Februar  12,  2013

Passwords and Login FailuresPasswords

• You may call 800-66-NERSC 24/7 to reset your password

• Change it at https://nim.nersc.gov

• Answer security questions in NIM, then you can reset it yourself at

https://nim.nersc.gov

20

Login Failures

• 3 or more consecutive login failures on a machine will disable your ability to log in

• Send e-mail to [email protected] or call 1-800-66-NERSC to reset your failure count

Page 21: Kirsten Fagnan NERSC User Services Februar  12,  2013

Data Resources

21

Page 22: Kirsten Fagnan NERSC User Services Februar  12,  2013

Data Storage Types• “Spinning Disk”

– Interactive access– I/O from compute jobs– “Home”, “Project”, “Scratch”, “Projectb”, “Global Scratch”– No on-node direct-attach disk at NERSC (except

PDSF,Genepool)• Archival Storage

– Permanent, long-term storage– Tapes, fronted by disk cache– “HPSS” (High Performance Storage System)

22

Page 23: Kirsten Fagnan NERSC User Services Februar  12,  2013

Home Directory•When you log in you are in your "Home" directory.•Permanent storage

– No automatic backups

•The full UNIX pathname is stored in the environment variable $HOME

–hopper04% echo $HOME

–/global/homes/r/ragerber

•$HOME is a global file system–You see all the same directories and files when you log in to any NERSC

computer.

•Your quota in $HOME is 40 GB and 1M inodes (files and directories). –Use “myquota” command to check your usage and quota

23

genepool04% echo $HOME/global/homes/k/kmfagnan

Page 24: Kirsten Fagnan NERSC User Services Februar  12,  2013

Scratch Directories•“Scratch” file systems are large, high-performance ”scratch” file

systems.

•Significant I/O from your compute jobs should be directed to $SCRATCH

•Each user has a personal directory referenced by $SCRATCH – on Genepool this points to /global/projectb/scratch/<username>

•Data in $SCRATCH is purged (12 weeks from last access)

•Always save data you want to keep to HPSS (see below)

•$SCRATCH is local on Edison and Hopper, but Carver and other systems use a global scratch file system. ($GSCRATCH points to global scratch on Hopper and Edison, as well as Genepool)

•Data in $SCRATCH is not backed up and could be lost if a file system fails.

24

Page 25: Kirsten Fagnan NERSC User Services Februar  12,  2013

Project Directories•All NERSC systems mount the NERSC global "Project”

file systems.

•Projectb is specific to the JGI, but is also accessible on Hopper and Carver.

•"Project directories” are created upon request for projects (groups of researchers) to store and share data.

• The default quota in /projectb is 5 TB.

•Data in /projectb/projectdirs is not purged and only 5TB is backed up. This may change in the future, but for long term storage, you should use the archive.

25

Page 26: Kirsten Fagnan NERSC User Services Februar  12,  2013

File System Access Summary

26

Page 27: Kirsten Fagnan NERSC User Services Februar  12,  2013

File System Summary

Page 28: Kirsten Fagnan NERSC User Services Februar  12,  2013

IO Tips• Use $SCRATCH for good IO performance • Write large chunks of data (MBs or more) at a time • Use a parallel IO library (e.g. HDF5)• Read/write to as few files as practical from your code (try

to avoid 1 file per MPI task)• Use $HOME to compile unless you have too many source

files or intermediate (*.o) files• Do not put more than a few 1,000s of files in a single

directory• Save any and everything important to HPSS

Page 29: Kirsten Fagnan NERSC User Services Februar  12,  2013

Archival Storage (HPSS)• For permanent, archival

storage• Permanent storage is

magnetic tape, disk cache is transient– 24PB data in >100M files written to

32k cartridges– Cartridges are loaded/unloaded into

tape drives by sophisticated library robotics

• Front-ending the tape subsystem is 150TB fast-access disk

29

• Hostname: archive.nersc.gov• Over 24 Petabyes of data stored• Data increasing by 1.7X per year• 120 M files stored• 150 TB disk cache• 8 STK robots• 44,000 tape slots• 44 PB maximum capacity today• Average data xfer rate: 100 MB/sec

Page 30: Kirsten Fagnan NERSC User Services Februar  12,  2013

Authentication• NERSC storage uses a token-based authentication method

– User places encrypted authentication token in ~/.netrc file at the top level of the home directory on the compute platform

• Authentication tokens can be generated in 2 ways:– Automatic – NERSC auth service:

• Log into any NERSC compute platform; Type “hsi”; Enter NERSC password

– Manual – https://nim.nersc.gov/ website• Under “Actions” dropdown, select “Generate HPSS Token”; Copy/paste content

into ~/.netrc; chmod 600 ~/.netrc

• Tokens are username and IP specific—must use NIM to generate a different token for use offsite

Page 31: Kirsten Fagnan NERSC User Services Februar  12,  2013

HPSS Clients• Parallel, threaded, high performance:

– HSI• Unix shell-like interface

– HTAR• Like Unix tar, for aggregation of small files

– PFTP• Parallel FTP

• Non-parallel:– FTP

• Ubiquitous, many free scripting utilities

• GridFTP interface (garchive)– Connect to other grid-enabled storage

systems

Page 32: Kirsten Fagnan NERSC User Services Februar  12,  2013

Archive Technologies, Continued…

• HPSS clients can emulate file system qualities– FTP-like interfaces can be deceiving: the archive is backed

by tape, robotics, and a single SQL database instance for metadata

– Operations that would be slow on a file system, e.g. lots of random IO, can be impractical on the archive

– It’s important to know how to store and retrieve data efficiently. (See http://www.nersc.gov/users/training/nersc-training-events/data-transfer-and-archiving/)

• HPSS does not stop you from making mistakes– It is possible to store data in such a way as to make it difficult to

retrieve– The archive has no batch system. Inefficient use affects others.

Page 33: Kirsten Fagnan NERSC User Services Februar  12,  2013

Avoid Common Mistakes• Don’t store many small files

– Make a tar archive first, or use htar• Don’t use recursively store or retrieve large directory trees• Don’t stream data via UNIX pipes

– HPSS can’t optimize transfers of unknown size• Don’t pre-stage data to disk cache

– May evict efficiently stored existing cache data• Avoid directories with many files

– Stresses HPSS database• Long-running transfers

– Can be error-prone– Keep to under 24 hours

• Use as few concurrent sessions as required– Limit of 15 in place

Page 34: Kirsten Fagnan NERSC User Services Februar  12,  2013

Hands-On• Use htar to save a directory tree to HPSS%cd $SCRATCH (or where you put the NewUser directory)

%htar –cvf NewUser.tar NewUser• See if the command worked%hsi ls –l NewUser.tar-rw-r----- 1 ragerber ccc 6232064 Sep 12 16:46 NewUser.tar

• Retreive the files%cd $SCRATCH2%htar –xvf NewUser.tar• Retreive the tar file%hsi get NewUser.tar

Page 35: Kirsten Fagnan NERSC User Services Februar  12,  2013

Data Transfer and Archiving• A NERSC Training Event• Good information:

http://www.nersc.gov/users/training/events/data-transfer-and-archiving/

35

Page 36: Kirsten Fagnan NERSC User Services Februar  12,  2013
Page 37: Kirsten Fagnan NERSC User Services Februar  12,  2013

Aside: Why Do You Care About Parallelism?

To learn more consider joining the HPC study group that meets on Friday from 12-1

37

Page 38: Kirsten Fagnan NERSC User Services Februar  12,  2013

Moore’s Law

38

2X transistors/Chip Every 1.5 yearsCalled “Moore’s Law”

Moore’s Law

Microprocessors have become smaller, denser, and more powerful.

Gordon Moore (co-founder of Intel) predicted in 1965 that the transistor density of semiconductor chips would double roughly every 18 months.

Slide source: Jack Dongarra

Page 39: Kirsten Fagnan NERSC User Services Februar  12,  2013

Power Density Limits Serial Performance

39

40048008

8080

8085

8086

286386

486Pentium®

P6

1

10

100

1000

10000

1970 1980 1990 2000 2010

Year

Pow

er D

ensi

ty (W

/cm

2 )

Hot Plate

NuclearReactor

RocketNozzle

Sun’sSurfaceSource: Patrick Gelsinger,

Shenkar Bokar, Intel

• High performance serial processors waste power- Speculation, dynamic dependence checking, etc. burn power- Implicit parallelism discovery

• More transistors, but not faster serial processors

• Concurrent systems are more power efficient – Dynamic power is

proportional to V2fC– Increasing frequency (f)

also increases supply voltage (V) cubic effect

– Increasing cores increases capacitance (C) but only linearly

– Save power by lowering clock speed

Page 40: Kirsten Fagnan NERSC User Services Februar  12,  2013

40

Revolution in Processors

• Chip density is continuing increase ~2x every 2 years• Clock speed is not• Number of processor cores may double instead• Power is under control, no longer growing

Page 41: Kirsten Fagnan NERSC User Services Februar  12,  2013

41

Moore’s Law reinterpreted• Number of cores per chip will double every two years• Clock speed will not increase (possibly decrease)• Need to deal with systems with millions of

concurrent threads• Need to deal with inter-chip parallelism as well as

intra-chip parallelism• Your take-away:

• Future performance increases in computing are going to come from exploiting parallelism in applications