scientific workflow systems for ... - university of chicago
TRANSCRIPT
Scientific Workflow Scientific Workflow Systems for 21st Century,Systems for 21st Century,New Bottle or New Wine?New Bottle or New Wine?
Ioan RaicuDistributed Systems LaboratoryComputer Science Department
University of Chicago
Collaborators: Yong Zhao, Microsoft CorporationIan Foster, University of Chicago and Argonne National Laboratory
Scientific Workflows (SWF 2008)July 8th, 2008
Workflow Systems
Describes computation components, ports and channels
Describes data and event flow
Coordinate the execution of the components
Scientific Workflow Systems for 21st Century, New Bottle or New Wine? 2
Functional MRI (fMRI)
Scientific Workflow Systems for 21st Century, New Bottle or New Wine? 3
• Wide range of analyses– Testing, interactive analysis,
production runs– Data mining– Parameter studies
B. Berriman, J. Good (Caltech)J. Jacob, D. Katz (JPL)
Scientific Workflow Systems for 21st Century, New Bottle or New Wine? 4
Molecular Dynamics
• Determination of free energies in aqueous solution– Antechamber – coordinates
Charmm solution– Charmm – solution– Charmm - free energy
5Scientific Workflow Systems for 21st Century, New Bottle or New Wine?
start
DOCK6Receptor
(1 per protein:defines pocket
to bind to)
ZINC3-D
structures
NAB scriptparameters
(defines flexibleresidues, #MDsteps)
BuildNABScript
NABScript
NABScript
Template
Amber prep:2. AmberizeReceptor4. perl: gen nabscript
FREDReceptor
(1 per protein:defines pocket
to bind to)
Manually prepDOCK6 rec file
Manually prepFRED rec file
1 protein(1MB)
6 GB2M
structures(6 GB)
DOCK6FRED ~4M x 60s x 1 cpu~60K cpu-hrs
PDBprotein
descriptions
Many Many Tasks:Identifying Potential Drug Targets
report ligands complexes
Amber Score:1. AmberizeLigand3. AmberizeComplex5. RunNABScript
end
Amber ~10K x 20m x 1 cpu~3K cpu-hrs
Select best ~500
~500 x 10hr x 100 cpu~500K cpu-hrsGCMC
Select best ~5KSelect best ~5K
For 1 target:4 million tasks
500,000 cpu-hrs(50 cpu-years)
Characterizing Scientific Workflows
• Inherit workflow system definition +…• Describe complex scientific procedures• Automate data derivation processes• High performance computing (HPC) to
improve throughput and performance• Provenance management and query
Scientific Workflow Systems for 21st Century, New Bottle or New Wine? 7
Characterizing Scientific Applications
• Increasing in scale and complexity– Wide inherent parallelism– Multiple successive stages– Wide range of number of tasks
• thousands to billions
Scientific Workflow Systems for 21st Century, New Bottle or New Wine? 8
• Potentially compute intensive– Widely varying task execution times
• 100s of ms to 10s of hours
• Potentially data intensive– I/O operation rates and aggregate I/O rates– Meta-data creation and modification– Significant data re-use
• Produced data is consumed by later stages
Many-Core Growth Rates
• Increasing attention toparallel chips– Many plans for cores
with “In-Order” execution
90 nm
65 nm
256 Cores
90 nm
65 nm
256 Cores
Slide 9
On-chip shared memoryFar faster to access on-chip memory than DRAM
Interesting challenges in synchronization (e.g. locking)
Inexpensive Low-Power Parallel ChipsAmazing amounts of computing very cheap
Slower (or same) sequential speed!2004 2006 2008 2010 2012 2014 2016 2018
45 nm 32
nm22 nm
16 nm 11
nm
8 nm2Cores
4Cores
8Cores
16Cores 32
Cores
64 Cores
128 Cores
2004 2006 2008 2010 2012 2014 2016 2018
45 nm 32
nm22 nm
16 nm 11
nm
8 nm2Cores
4Cores
8Cores
16Cores 32
Cores
64 Cores
128 Cores
Pat Helland, Microsoft, The Irresistible Forces Meet the Movable Objects, November 9th, 2007
What will we do with 1+ Exaflopsand 1M+ cores?
1) Tackle Bigger and BiggerProblems
ComputationalS i ti t
Scientific Workflow Systems for 21st Century, New Bottle or New Wine? 11
Scientistas
Hero
2) Tackle Increasingly ComplexProblems
ComputationalScientist
as Logistics
Scientific Workflow Systems for 21st Century, New Bottle or New Wine? 12
LogisticsOfficer
“More Complex Problems”
• Ensemble runs to quantify climate model uncertainty• Identify potential drug targets by screening a database of
ligand structures against target proteins• Study economic model sensitivity to parameters
A l t b l d t t f ti
Scientific Workflow Systems for 21st Century, New Bottle or New Wine?13
• Analyze turbulence dataset from many perspectives• Perform numerical optimization to determine optimal resource
assignment in energy problems• Mine collection of data from advanced light sources• Construct databases of computed properties of chemical
compounds• Analyze data from the Large Hadron Collider• Analyze log data from 100,000-node parallel computations
Programming Model Issues
• Massive task parallelism• Massive data parallelism• Integrating black box applications• Complex task dependencies (task graphs)
Scientific Workflow Systems for 21st Century, New Bottle or New Wine?14
• Complex task dependencies (task graphs)• Failure, and other execution management issues• Data management: input, intermediate, output• Dynamic computations (task graphs)• Dynamic data access to large, diverse datasets• Long-running computations• Documenting provenance of data products
Problem Types
Inputdatasize
Hi
Med
Dataanalysis,mining
Big data and many tasks
Scientific Workflow Systems for 21st Century, New Bottle or New Wine? 15Number of tasks
1 1K 1M
Med
Lo
HeroicMPItasks
Many loosely coupled tasks
An Incomplete and Simplistic View of Programming Models
and Tools
Many Tasks Much Data
Single task, modest dataMPI, etc., etc., etc.
Many TasksDAGMan+Pegasus
Karajan+Swift+Falkon
Much DataMapReduce/Hadoop
Dryad
Complex Tasks, Much DataDryad, Pig, Sawzall
Swift+Falkon (using data diffusion)
Scientific Workflow Systems for 21st Century, New Bottle or New Wine?
Major Challenges to Large Scale Scientific Computation
• Managing heterogeneous scientific data– Idiosyncratic layouts & formats (file sys, db, spreadsheet, XML,
etc…) • Describing complex science problems• Coordinating distributed diverse computation proceduresg p p
– Executables, scripts, Web services• Scheduling & executing numerous tasks reliably and efficiently
– Large quantity of data (Petabytes/year)– Large number of parallel/dependent tasks (103~106 tasks)
• Organizing, archiving and tracking– Datasets, procedures, workflows, provenance
18Scientific Workflow Systems for 21st Century, New Bottle or New Wine?
Sw
ift
BP
EL
XP
DL
MS
Wo
rkflo
wFu
nd
atio
n
DA
GM
an
Taven
a
Tria
na
Kep
ler
Map
red
uce
Sta
r-P
Scales to Grids ++ - - - ++ - - - + +
Existing and emerging workflow technologies
19Scientific Workflow Systems for 21st Century, New Bottle or New Wine?
++ ++ + +
Typing ++ ++ ++ ++ - - - + - +
Iteration ++ -/+ - + - - - + + +
Scripting ++ - - + + + - - - ++
Dataset Mapping + - - - - - - - - -
Service Interop + - + - - - - + - -
Subflow/comp. + - + + - - + + - +
Provenance + - - + - + - + - -
Open source + + + - + + + + - -
Virtual Node(s)Abstractcomputation
S iftS i t
Specification Execution
Virtual Node(s)
file1
Scheduling
Execution Engine(Karajan w/
Swift Runtime)
Swift Architecture
Provisioning
FalkonResource
Provisioner
SwiftScript
Virtual DataCatalog
SwiftScriptCompiler
Provenancedata
ProvenancedataProvenance
collector
launcher
launcher
file2
file3
AppF1
AppF2
Swift runtimecallouts
CC CC
Status reportingAmazon
EC2
20Scientific Workflow Systems for 21st Century, New Bottle or New Wine?
0 1800 3600 5400 7200 9000 10800 12600 144001
Time (sec)
• 244 molecules 20497 jobs• 15091 seconds on 216 CPUs 867.1 CPU hours• Efficiency: 99.8%• Speedup: 206.9x 8.2x faster than GRAM/PBS• 50 molecules w/ GRAM (4201 jobs) 25.3 speedup
MolDyn Application
Scientific Workflow Systems for 21st Century, New Bottle or New Wine? 2121
1100120013001400150016001700180019001
1000111001120011300114001150011600117001180011900120001
Task
ID
waitQueueTime execTime resultsQueueTime
Managing 120K CPUs
High-speed local disk
Falkon
22Scientific Workflow Systems for 21st Century, New Bottle or New Wine?
Slower shared storage
MARS Economic Modeling on IBM BG/P
• CPU Cores: 2048• Tasks: 49152• Micro-tasks: 7077888• Elapsed time: 1601 secs• CPU Hours: 894
Scientific Workflow Systems for 21st Century, New Bottle or New Wine? 23
• CPU Hours: 894• Speedup: 1993X (ideal 2048)• Efficiency: 97.3%
MARS Economic Modeling on IBM BG/P (128K CPUs)
3500
4000
4500
5000
800000
1000000
ec)
s
ProcessorsActive TasksTasks CompletedThroughput (tasks/sec)
• CPU Cores: 130816• Tasks: 1048576• Elapsed time: 2483 secs• CPU Years: 9.3
Speedup: 126892X (ideal 130816)
Scientific Workflow Systems for 21st Century, New Bottle or New Wine? 24
0
500
1000
1500
2000
2500
3000
3500
0
200000
400000
600000
Thro
ughp
ut (t
asks
/se
Task
s C
ompl
eted
Num
ber o
f Pro
cess
ors
Time (sec)
Speedup: 126892X (ideal 130816)Efficiency: 97%
Scientific Workflow Systems for 21st Century, New Bottle or New Wine? 25
Many Many Tasks:Identifying Potential Drug Targets
2M+ ligandsProtein xtarget(s)
(Mike Kubal, Benoit Roux, and others)26Scientific Workflow Systems for 21st Century, New Bottle or New Wine?
DOCK on SiCortex
• CPU cores: 5760• Tasks: 92160• Elapsed time: 12821 sec• Compute time: 1.94 CPU years
Scientific Workflow Systems for 21st Century, New Bottle or New Wine? 27
• Average task time: 660.3 sec• Speedup: 5650X (ideal 5760)• Efficiency: 98.2%
DOCK on the BG/P
CPU cores: 118784Tasks: 934803Elapsed time: 2.01 hoursCompute time: 21.43 CPU yearsAverage task time: 667 secRelative Efficiency: 99 7%
28
Relative Efficiency: 99.7%(from 16 to 32 racks)Utilization: • Sustained: 99.6%• Overall: 78.3%
Time (secs)
Scientific Workflow Systems for 21st Century, New Bottle or New Wine?
Scientific Workflow Systems for 21st Century, New Bottle or New Wine? 29
Support for Data Intensive Applications (Falkon and Data Diffusion)
• Resource acquired in response to demand
• Data and applications diffuse from archival storage to newly acquired resources
text
Task DispatcherData-Aware Scheduler Persistent Storage
Shared File System
Idle Resources
Scientific Workflow Systems for 21st Century, New Bottle or New Wine? 30
• Resource “caching” allows faster responses to subsequent requests – Cache Eviction Strategies:
RANDOM, FIFO, LRU, LFU• Resources are released
when demand drops
Shared File System
Provisioned Resources
AstroPortal Stacking Servicewith Data Diffusion
• Aggregate throughput:– 39Gb/s– 10X higher than GPFS
• Reduced load on GPFS– 0.49Gb/s
1/10 of the original load20
25
30
35
40
45
50
te T
hrou
ghpu
t (G
b/s)
Data Diffusion Throughput LocalData Diffusion Throughput Cache-to-CacheData Diffusion Throughput GPFSGPFS Throughput (FIT)GPFS Throughput (GZ)
Scientific Workflow Systems for 21st Century, New Bottle or New Wine? 310
200
400
600
800
1000
1200
1400
1600
1800
2000
2 4 8 16 32 64 128
Number of CPUs
Tim
e (m
s) p
er s
tack
per
CPU
Data Diffusion (GZ)Data Diffusion (FIT)GPFS (GZ)GPFS (FIT)
High data locality– Near perfect scalability
– 1/10 of the original load
0
5
10
15
1 1.38 2 3 4 5 10 20 30Locality
Agg
rega
t
Data Diffusion:Data-Intensive Workload
• 250K tasks on 128 processors– 10MB read, 10ms compute
• Comparing GPFS with data diffusion– 5011 sec vs. 1427 sec (ideal is 1415 sec)
1000 11000
Scientific Workflow Systems for 21st Century, New Bottle or New Wine? 32
0.001
0.01
0.1
1
10
1000
120
240
360
480
600
720
840
960
1080
1200
1320
1440
Time (sec)
Num
ber o
f Nod
esA
ggre
gate
Thr
ough
put (
Gb/
s)Th
roug
hput
per
Nod
e (M
B/s
)Q
ueue
Len
gth
(x1K
)
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
Cac
he H
it/M
iss
%
Cache Hit Local % Cache Hit Global % Cache Miss %Ideal Throughput (Gb/s) Throughput (Gb/s) Wait Queue LengthNumber of Nodes
0.001
0.01
0.1
1
10
100
030
060
090
012
0015
0018
0021
0024
0027
0030
0033
0036
0039
0042
0045
0048
0051
00
Time (sec)
Num
ber o
f Nod
esTh
roug
hput
(Gb/
s)Q
ueue
Len
gth
(x1K
)
Ideal Throughput (Gb/s) Throughput (Gb/s) Wait Queue Length Number of Nodes
Hadoop vs. Swift
• Classic benchmarks for MapReduce– Word Count– Sort
• Swift performs similar or better than Hadoop
Scientific Workflow Systems for 21st Century, New Bottle or New Wine? 33
p p(on 32 processors)
Sort
4285
733
25
83
512
1
10
100
1000
10000
10MB 100MB 1000MBData Size
Tim
e (s
ec)
Swift+Falkon
Hadoop
Word Count
221
11431795
863
46887860
1
10
100
1000
10000
75MB 350MB 703MBData Size
Tim
e (s
ec)
Swift+PBSHadoop
Mythbusting
• Embarrassingly Happily parallel apps are trivial to run– Logistical problems can be tremendous
• Loosely coupled apps do not require “supercomputers”– Total computational requirements can be enormous– Individual tasks may be tightly coupled
W kl d f tl i l l t f I/O
Scientific Workflow Systems for 21st Century, New Bottle or New Wine? 34
– Workloads frequently involve large amounts of I/O– Make use of idle resources from “supercomputers” via backfilling – Costs to run “supercomputers” per FLOP is among the best
• BG/P: 0.35 gigaflops/watt (higher is better)• SiCortex: 0.32 gigaflops/watt• BG/L: 0.23 gigaflops/watt• x86-based HPC systems: an order of magnitude lower
• Loosely coupled apps do not require specialized system software• Shared file systems are good for all applications
– They don’t scale proportionally with the compute resources– Data intensive applications don’t perform and scale well
Features Scientific Workflow Systems should Have!
• Parallelism– Support for both explicit and implicit parallelism
• Performance and Scalability– Million to billions of tasks– Handle 100s~1000s of tasks/sec
35Scientific Workflow Systems for 21st Century, New Bottle or New Wine?
– Handle 100s 1000s of tasks/sec
• Data management– Reduce reliance on shared file systems– Scale with processing power– Data-aware scheduling
• Reliability– Self healing– Efficient and scalable monitoring
• Provenance
Solutions(we have experience with)
• Falkon– A Fast and Light-weight tasK executiON framework– Globus Incubator Project– http://dev.globus.org/wiki/Incubator/Falkon
• SwiftP ll l i t l f id d li bl ifi ti ti
Scientific Workflow Systems for 21st Century, New Bottle or New Wine? 36
– Parallel programming tool for rapid and reliable specification, execution, and management of large-scale science workflows
– http://www.ci.uchicago.edu/swift/index.php• Environments:
– Clusters: TeraPort (TP)– Grids: Open Science Grid (OSG), TeraGrid (TG)– Specialized large machines: SiCortex 5732– Supercomputers: IBM BlueGene/P (BG/P)
More Information
• More information: – Personal research page: http://people.cs.uchicago.edu/~iraicu/– Falkon: http://dev.globus.org/wiki/Incubator/Falkon– Swift: http://www.ci.uchicago.edu/swift/index.php
• Collaborators (relevant to this proposal):– Ian Foster, The University of Chicago & Argonne National Laboratory– Alex Szalay, The Johns Hopkins University
Y Zh Mi ft– Yong Zhao, Microsoft– Mike Wilde, Computation Institute, University of Chicago & Argonne National
Laboratory – Catalin Dumitrescu, Fermi National Laboratory– Zhao Zhang, The University of Chicago– Jerry C. Yan, NASA, Ames Research Center
• Funding:– NASA: Ames Research Center, Graduate Student Research Program (GSRP)– DOE: Mathematical, Information, and Computational Sciences Division
subprogram of the Office of Advanced Scientific Computing Research, Office of Science, U.S. Dept. of Energy
– NSF: TeraGrid
Scientific Workflow Systems for 21st Century, New Bottle or New Wine? 37
Scientific Workflow Systems for 21st Century, New Bottle or New Wine? 38
Handling MegajobsBOF at SC08
• More and more people need to run thousands to millions of closely related jobs that are associated with individual projects. Scientists seek convenient means to specify and manage many jobs, arranging inputs, aggregating outputs, identifying successful and failed jobs and repairing failures. System administrators seek methods to process extraordinary numbers of jobs for multiple users without overwhelming queuing systems or disrupting fair-share usage policies And grid developers are producing a new generationshare usage policies. And, grid developers are producing a new generation of queuing and scheduling systems as well as auxiliary systems for use with existing queuing and scheduling systems. This Birds-of-feather session provides a venue for the exchange of information about processing large numbers of jobs. Short presentations of an invited sample of projects will be followed by discussion.
• For more information, contact:– Marlon Pierce: [email protected]– Dick Repasky: [email protected]– Ioan Raicu: [email protected]
Scientific Workflow Systems for 21st Century, New Bottle or New Wine? 39