suragrid update sura bof @ internet2 smm april 23, 2007 mary fran yafchak, [email protected] sura it...

21
SURAgrid Update SURA BoF @ Internet2 SMM April 23, 2007 Mary Fran Yafchak, [email protected] SURA IT Program Coordinator, SURAgrid project manager

Upload: claude-richardson

Post on 26-Dec-2015

215 views

Category:

Documents


1 download

TRANSCRIPT

SURAgrid UpdateSURA BoF @ Internet2 SMMApril 23, 2007

Mary Fran Yafchak, [email protected] IT Program Coordinator, SURAgrid project manager

SURAgrid Vision

SURA regional development:• Develop & manage partnership relations• Facilitate collaborative project development• Orchestrate centralized services & support• Foster and catalyze application development• Develop training & education (user, admin)• Other…(Community-driven, over time…)

Project-Specific View

SURAgrid Resources

SURAgrid Resourcesand Applications

“MySURAgrid” View

Sample User Portals

Industry PartnerCoop Resources

(e.g. IBM partnership)

InstitutionalResources

(e.g. Current participants)

Gateways to National Cyberinfrastructure

(e.g. Teragrid)

Other externallyfunded resources

(e.g. group proposals)

VO or ProjectResources

(e.g, SCOOP)

Heterogeneous Environment to Meet Diverse User Needs

SURAgrid

Project-specific tools

SURAgrid Growth & “Mix” University of Alabama at

Birmingham University of Alabama in Huntsville University of Florida Georgia State University University of Michigan University of Southern California Texas Advanced Computing Center University of Virginia

As of April 2004 (25%)

(Key: SURA member; not SURA member)

SURAgrid Growth & “Mix” University of Alabama at Birmingham University of Alabama in Huntsville University of Florida Georgia State University University of Michigan University of Southern California Texas Advanced Computing Center University of Virginia Tulane University University of Arkansas George Mason University University of South Carolina Great Plains Network Louisiana State Mississippi Center for SuperComputing

Research

Texas Tech University of Louisiana at Lafayette North Carolina State University Texas A&M University

As of April 2005 (21%)

(Key: SURA member; not SURA member)

SURAgrid Growth & “Mix” University of Alabama at Birmingham University of Alabama in Huntsville University of Florida Georgia State University University of Michigan University of Southern California Texas Advanced Computing Center University of Virginia Tulane University University of Arkansas George Mason University University of South Carolina Great Plains Network Louisiana State Mississippi Center for SuperComputing

Research

Texas Tech University of Louisiana at Lafayette North Carolina State University Texas A&M University University of North Carolina, Charlotte University of Kentucky Old Dominion University Vanderbilt University Bowie State University Louisiana Tech University University of Maryland

As of April 2006 (23%)

(Key: SURA member; not SURA member)

SURAgrid Growth & “Mix” University of Alabama at Birmingham University of Alabama in Huntsville University of Florida Georgia State University University of Michigan University of Southern California Texas Advanced Computing Center University of Virginia Tulane University University of Arkansas George Mason University University of South Carolina Great Plains Network Louisiana State Mississippi Center for

SuperComputing Research

Texas Tech University of Louisiana at Lafayette North Carolina State University Texas A&M University University of North Carolina,

Charlotte University of Kentucky Old Dominion University Vanderbilt University Bowie State University Louisiana Tech University University of Maryland Kennesaw State University Clemson University Wake Forest University

As of April 2007 (27%)Potential with talks in-progress: 33%

(Key: SURA member; not SURA member)

GSU

UAB

USC

ULL

TAMU UFL

MCSR

Tulane

UArk

TTUSC

LSU

GPN

ODU

UNCC

NCState

Vanderbilt

UAH

UMich UKY

UVA

TACC

GMU

SURAgrid Participants(As of March 2007)

= SURA Member = Resources on the grid

UMD

Bowie State

LATechKennesaw

State

Clemson

= non-SURA Participant

Wake Forest

And new SURAgrid staff too!

A warm welcome to… Linda Akli, Grid Application & Outreach

Specialist (as of January 2007)

Mary Trauner, Contractor in resource deployment & support (as of last week!)

SURAgrid Infrastructure Specialist (coming soon - interviews in progress…)

SURAgrid Application Growth Five teams showcased their app at I2 FMM in Dec 07

– GSU: Multiple Genome Alignment on the Grid– UAB: Dynamic BLAST– ODU: Bioelectric Simulator for Whole Body Tissues– NCState: Simulation-Optimization for Threat Management in

Urban Water Systems– UNC: Storm Surge Modeling with ADCIRC

Documentation– Application “glossies” (see handouts)– Running Your Application on SURAgrid, v2

• http://www.sura.org/programs/docs/RunAppsSURAgridV2.pdf And more applications in the pipeline (see handout)

More on SCOOP & SURAgrid Current collaboration

– SURA IT staff in SCOOP planning & communications– Active deployment on SURAgrid resources

• Supported individual model runs in 2006 hurricane season• Planning to support integrated model runs in 2007

Future collaboration?– SURA IT & SCOOP management seeking to increase

resources for SCOOP-SURAgrid integration– Areas of potential co-development:

• Event-driven resource scheduling• Data storage, discovery, redundancy, etc. (“data grid”)• Secure multi-level access (Integrated portals, tiered PKI?)• Access to computation & facilities for high quality visualization

SURAgrid Resource growth Contributors about the same but nature is changing…

– Several test resources being replaced by larger resources, or those from production environments

– Several institutions now contributing multiple resources New additions

– GSU’s IBM P575 (first from the SURAgrid-IBM partnership)• LSU & TAMU integrations pending…

– LSU’s SuperMike• See http://www.hpcwire.com/hpc/1320723.html

– Production cluster from UMich (also supports ATLAS)• See https://hep.pa.msu.edu/twiki/bin/view/AGLT2

Total combined capacity as of March 2007: 1991 CPUs for total of approximately 10.5 teraflops, with 4.6 terabytes of storage.

Infrastructure Development Updates to SURAgrid portal (maintained by TACC):

– Improvements in user documentation – Enhanced system detail & resources “at a glance”– Test scripts at the service level – See Documentation & Resource Monitor at gridportal.sura.org

Access management (developed by UVA):– Tiered-PKI being designed to provide multiple levels of

assurance within SURAgrid– Plans underway for most critical but affordable components

while seeking funding for the rest– SURAgrid CA (at current LoA), strong authentication for site

administrators, foundational policy documentation

Infrastructure Development

New! SURAgrid Accounting Working Group to address the immediate need for grid-wide views of resource and application activity– Participants: UAB, GMU, GSU, LSU, UMich, ODU,

TAMU, TTU; SURA facilitating

New! SURAgrid “stack packaging” task force– TTU leading, with participation from GSU, GMU,

LSU, UAB

And with growth…we need things like: More formal governance process Increased need to find the means to expand

& sustain the effort

Governance & Decision-Making

SURAgrid Project Planning Working Group– Established at Sept 2006 In-Person Meeting to develop

governance options for SURAgrid– Participants:

• Linda Akli, SURA• Gary Crane, SURA• Steve Johnson, Texas A&M University• Sandi Redman, University of Alabama in Huntsville• Don Riley, University of Maryland & SURA IT Fellow• Mike Sachon, Old Dominium University• Srikanth Sastry, Texas A&M University• Art Vandenberg, Georgia State University• Mary Fran Yafchak, SURA

Governance Overview To date SURAgrid has used consensus-based decision

making with SURA facilitating Maturity & investment requires more formal structure

– Ensure those investing have appropriate role in governance– Support sustainable growth & active participation

Suggested initial classes of membership– Contributing Member: HE or related org contributing significant

resources to advance SURAgrid infrastructure• SURA is contributing member by definition

– Participating Member: HE or related org participating in SURAgrid activities other than Contributing Member

– Partnership Member: Entity (org, commercial, non-HE…) with strategic relationship with SURAgrid

Proposed Governance Structure

Formation of SURAgrid Governance Committee – Primary governing body, elected by SURAgrid Contributing

Members – Will act on behalf of SURAgrid community

• Guidance, Facilitation, Reporting

Initial SURAgrid Governance Committee will have 9 members: – 8 elected by current SURAgrid Participating Organizations, 1

appointed by SURA– Current SURAgrid Participating Organizations defined as all

those participating as of December 31, 2006

Transitioning to New Governance SURAgrid Participating Organization designate a

SURAgrid lead representative – Done New governance process/structure approved by SURA

IT Steering Group – Done New Governance process/structure approved by vote

of SURAgrid Participating leads – Done Call for nominations for SURAgrid Governance

Committee candidates – Done Nominations will be accepted through midnight April 28

– In progress Election of SURAgrid Governance Committee members

is expected to be completed by May 12

Sustaining the effort, Part 1 Continuing to seek external funding for SURAgrid

through collaborative proposal development – SURA-led NSF STCI (Strategic Technologies for

Cyberinfrastructure) proposal (under review)

• Expanded grid-wide operations & support, Portal maintenance & development, Development of two-tiered PKI (Public Key Infrastructure)

• Submission included 9 letters of support from other SURAgrid sites as well as Teragrid, SCOOP & NCSA

– SURAgrid as grid service provider in NSF CCLI (Course, Curriculum & Laboratory Improvement) proposal led by UNC-C (Barry Wilkinson) & UArk (Amy Apon) (under review)

• Other partners: LSU, GMU, Lewis & Clark University, UNC Wilmington, Southern University & Howard University

Sustaining the effort, Part 2 SURAgrid included as participant in LSU’s Nov 06 submission

to NSF 06-599 “High-Performance Computing for Science and Engineering

SURAgrid was also a partner in LSU’s Feb 07 NSF proposal submission to the NSF 06-573 “Leadership-Class System Acquisition - Creating a Petascale Computing Environment for Science and Engineering” program.

SURA staff have also been working with John Connolly (Director of the University of Kentucky Center for Computational Sciences) and several IT leaders from SURA’s EPSCoR community to explore the development of a SURA program to leverage SURAgrid and SURAgrid’s Corporate Partners program to create a cyberinfrastructure program targeted at the SURA EPSCoR community.

Now to Gary (Whew!) for more detail on corporate partnerships…