scidac - accelconf.web.cern.ch

37
Accelerator Modelling under SciDAC: Meeting the Challenges of Next- Generation Accelerator Design, Analysis, and Optimization Panagiotis Spentzouris Fermi National Accelerator Laboratory ICAP 2006 October 5, 2006 Chamonix

Upload: others

Post on 03-Jan-2022

3 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: SciDAC - accelconf.web.cern.ch

Accelerator Modelling under SciDAC:

Meeting the Challenges of Next-Generation Accelerator Design,

Analysis, and Optimization

Panagiotis Spentzouris

Fermi National Accelerator Laboratory

ICAP 2006

October 5, 2006

Chamonix

Page 2: SciDAC - accelconf.web.cern.ch

OUTLINE

• SciDAC1: objectives and development approach

• SciDAC1 accomplishments: impact to present and future facilities

• SciDAC2: planning the future

Page 3: SciDAC - accelconf.web.cern.ch

and BTW,

SciDAC: Scientific Development through Advanced Computing.

SciDAC1: 2001-2006SciDAC2: 2007-2012 (?)

Page 4: SciDAC - accelconf.web.cern.ch

The SciDAC1 program

• Develop scientific applications to effectively take advantage of terascalecomputing, by — Creating a new generation of scientific

simulation codes— Creating the mathematical and computing

systems software to enable these scientific simulation codes to use terascale computers

— Creating a distributed science software infrastructure to enable scientists to effectively utilize these codes.

Accelerator modeling: emphasis on building teams of computer scientists, and computational and machine

physicists

Page 5: SciDAC - accelconf.web.cern.ch

SciDAC Accelerator Science & Technology (AST) Project

Goals: Develop & apply an advanced, comprehensive, high-performance simulation environment to solve challenging problems and to enable new discoveries in accelerator science & technology

Participants:Labs: LBNL, SLAC, FNAL, LANL, BNL, SNLUniversities: Stanford, UCLA, USC, UCD, UMdSmall business: Tech-XCollaborations w/ applied math and CS researchers

Sponsors: DOE/SC HEP (formerly HENP) in collaboration with ASCR

Page 6: SciDAC - accelconf.web.cern.ch

Collaborations w/ Applied Math & Comp. SciSciDAC Integrated Software Infrastructure Centers

Linear solvers, eigensolversPoisson solversAMRMeshingParallel PIC methodsPerformance optimizationStatistical methodsVisualization

Page 7: SciDAC - accelconf.web.cern.ch

AST Thrust Areas & Codes

Beam Dynamics (BD)BeamBeam3D, IMPACT-Z, IMPACT-T, ML/I, Synergia

Electromagnetics (EM)Omega3P, Tau3P, S3P, T3P, Track3P

Advanced Accelerators (AA)OSIRIS, VORPAL, QuickPIC, UPIC

Emphasis of this talk will be on BD and AA

Page 8: SciDAC - accelconf.web.cern.ch

SciDAC/AST codes have been applied to present & futureprojects

TevatronLHCNLCILCPEP-IIFNAL BoosterFNAL Main InjectorL’OASIS LWFA experimentsSLAC PWFA experimentsPlasma afterburner designRHICRIASNSLCLSPhotoinjector designAdvanced streak camera designCERN SPSJPARC commissioningFERMI designInt’l code benchmarking @ CERN PS

Page 9: SciDAC - accelconf.web.cern.ch

12 SciDAC project related talks in ICAP06!

J. Cary, “High Performance Self-Consistent EM modeling of beams”, MOAPMP02 J.-F. Ostiguy, “CHEF: A Framework for Accelerator Optics and Simulation”, TUAPMP02 R. Ryne, “Recent Progress on the MaryLie/IMPACT Beam Dynamics Code”, TUAPMP03 A. Candell, “Parallel Higher-Order Finite Element Method for Accurate Field Computations in Wakefield And PIC Simulations”, WEMPMP03 I. Pogorelov, “Recent Developments in IMPACT and Application for Future Light Sources”, WEPPP01 J. Qiang, “Recent Improvements to the IMPACT-T Parallel Particle Tracking Code”, WEPPP02

C. E. Mitchell, “Computation of Transfer Maps from Surface Data with Applications to Wigglers”, WEPPP08 D. L. Bruhwiler, “High-Order Algorithms for Simulation of Laser Wakefield Accelerators”, WESEPP03 “D. L. Bruhwiler Parallel Simulation of Coulomb Collisions for High-Energy Electron Cooling Systems”, WEA1MP01 J-L Vay, “Self-Consistent Simulations of High-Intensity Beams and E-Clouds with WARP/POSINST”, WEA3MP02 A. Kabel, “Accelerating Cavity Design for the International Linear Collider”, WEA4IS01 A. Adelmann, “H5Part: A Portable High Performance Parallel Data Interface for Electromagnetics Simulations”, THM1MP01

SciDAC2 codes

Codes used in SciDACbut effort not SciDAC

funded

SciDAC codes

Page 10: SciDAC - accelconf.web.cern.ch

SciDAC AST codes: notable “firsts”

First 100M strong-strong colliding beam simulation for LHC (BeamBeam3D; J. Qiang)

First multi-bunch, multi-turn injection simulation from linac-to-booster w/ self-consistent 3D space charge (Synergia; J. Amundson and P. Spentzouris)

First 100M simulation of a linac for an x-ray light source w/ self-consistent 3D space charge (IMPACT-Z; I. Pogorelov, J. Qiang)

First self-consistent electromagnetic simulation of an intense beam in an ILC 'crab' cavity (VORPAL; J.R. Cary, C. Nieter & VORPAL team)

First 3D simulation of a 1TeV Afterburner stage (QuickPIC; C.K.Huang et al.)C.K. Huang received the 2007 Nicholas Metropolis award

First 3D simulation of a GeV LWFA stage (OSIRIS; F.S.Tsung, W.Lu, M. Tzoufras et al.)

Page 11: SciDAC - accelconf.web.cern.ch

Selected AST project code descriptions and

applications

Page 12: SciDAC - accelconf.web.cern.ch

BeamBeam3D• Multi-model parallel PIC code for simulating colliding beams

(weak-strong,strong-strong,head-on,crossing angle, long range)

• Applied to Tevatron,

LHC, PEP-II, RHIC• Features: Integrated and non-uniform grid Green function,

multi-slice/multi-bunch/multi-IP, impedance model

7.181024

14.77256

4964

18816

Execution time/turn (sec)

# of procs

Scaling at NERSC SP3: weak-strong model (100Mp, 512x512x32 grid, 4 slices)

PEP-II simulations show need for multi-slice modeling for accurate luminosity calculation. (J. Qiang/LBNL, Y. Cai/SLAC, K. Ohmi/KEK)

20 slices1 slice

BramBeam3d

0

0.5

1

1.5

2

2.5

0 500 1000 1500

Number of processors

log

(sec

s/tu

rn)

Page 13: SciDAC - accelconf.web.cern.ch

Simulation of beam-beam effects in VEPP-II:Comparing matrix model, BeamBeam3D, expt

Beam-beam code validation comparingwith VEPP-II data

2-bunch coherent spectrum for ξ=0.008. Shown are the beam-beam σ and π modes and one

synchro-mode.

E. Stern, A. Valisev (FNAL), J. Qiang (LBNL)

1e-06

1e-05

1e-04

0.001

0.01

0.1

0.07 0.08 0.09 0.1 0.11 0.12 0.13 0.14 0.15

po

we

r (a

rbitr

ary

un

its)

tune

sigma modesynchro mode 1

beam-beam pi mode

Page 14: SciDAC - accelconf.web.cern.ch

Application of QuickPIC to Electron-Cloud :A few 10000’s km beam-cloud interaction

Ali Ghalam, B. Feng, T. Katsouleas (USC); W. B. Mori, C. Huang, V. Decyk, (UCLA); G. Rumolo, F. Zimmermann, E. Benedetto (CERN); P. Spentzouris (FNAL)

+

-

SynchrotronRadiation

SecondaryEmission

Circularmachine

Positron BunchQuickPIC adapted to circular problem

•Magnetic field•Non-neutral plasma•Lattice effects

•Can now model 100 msecs

beam cloud

Illustrates how multidisciplinary collaboration can lead to unexpected benefits: A laser/plasma code applied to LHC and FNAL MI.

Beam Bunch

Page 15: SciDAC - accelconf.web.cern.ch

QuickPICQuickPIC LHC modeling: Snap shots of Beam Evolution LHC modeling: Snap shots of Beam Evolution

Z

y

T=0 After 250 turnsHead-Tail Dephasing

After 500 turnsEmittance Growth

After 1000 turns After 2000 turns

0.8

0.9

1

1.1

1.2

1.3

1.4

1

74 147

220

293

366

439

512

585

658

731

804

877

950

1023

1096

1169

1242

1315

1388

1461

1534

1607

1680

No. of Turns

Siz

e(m

m)

Single kick not suffcicient: A. Z. Ghalam, T. Katsouleas(USC)

G. Rumolo, F.Zimmermann(CERN)

C. Huang, V. Decyk, W.Mori(UCLA)

Page 16: SciDAC - accelconf.web.cern.ch

IMPACT: Integrated-Map & Particle Accelerator Tracking

• Code suite includes IMPACT-Z, IMPACT-T parallel PIC codes

• Originally for ion linacs; major enhancements under SciDAC for electron linacs, photoinjectors, …

• Recent enhancements—high aspect ratio Poisson solver—binning for large energy spread—multi-charge state (RIA)—wakes, 1D CSR

Au Photocathode-10 kV

Magnetic lensMCP

Sample in transmissionor reflection

Tim

e

Space

RecordedStreak image

View from side

Plates parallel to timing slit

Sw

eep

0.25 mm

Phosphor

CCD

• Applied to SNS, RIA, JPARC, Fermi@Elettra

• photoinjectors @ ANL, BNL, Cornell, FNAL/NIU, JLAB, LBNL, SLAC/LCLS

Emission from nano-needle tip

• Next-gen streak camera design

Page 17: SciDAC - accelconf.web.cern.ch

LCLS photoinjector emittance evolution (J. Qiang/LBNL, C. Limbourg/SLAC)

experiment simulation

Ion beam formation & transport from RIA ECR ion source (J. Qiang)

JPARC commissioning, horizontal phase space, simulation vs. expt (M. Ikegami/KEK)

Example IMPACT applications: HEP, NP, BES

Beam size vs z in the FNAL/NICADD photoinjector: simulation & experiment (C. Bohn/NIU, F. Piot/FNAL)

Page 18: SciDAC - accelconf.web.cern.ch

~ ~beta=0.52 beta=0.70

propagate envelopes to here

compute matched beam in this period

adjust 4 quad strengths + 3 rf cavity parameters to fit 6 envelope functions (2 each for x,y,z) + Efinal

#13period: #14 #15 #16

MaryLie/IMPACT (ML/I)• ML/I: hybrid code combining MaryLie 5th order magnetic

optics with IMPACT parallel PIC + new capabilities—Embeds operator splitting for all thick elements—New modules (wakefields, soft-edge magnet models, …)

• Multiple-physics, multi-purpose—Particle tracking, envelope tracking, map production/analysis—Fitting/optimizing, e.g. zeroing 3rd order while minimizing 5th

—Designing matching sections, e.g. superconducting linacs

Page 19: SciDAC - accelconf.web.cern.ch

Following the SciDAC model, AST codes are developed by large, multi-disciplinary teams

Example: MaryLie/IMPACT

Page 20: SciDAC - accelconf.web.cern.ch

ML/I applications

ILC damping ring simulations showing emittance growth w/ linear (left), nonlinear (right) space-charge models. M. Venturini, LBNL. These simulations

contributed to the ILC-DR design selection.

Page 21: SciDAC - accelconf.web.cern.ch

Synergia• Multi-language, extensible, parallel framework

• Incorporates multi-physics; state-of-the-art numerical libraries,

solvers, physics modules

3D space-charge, impedance, arbitrary

order lie maps, …

Page 22: SciDAC - accelconf.web.cern.ch

− Multi-bunch capability

− Multiple Poisson solvers

− IMPACT, multigrid (PETSc)

− Multi-turn injection

− Ramping rf and magnet modeling

− Active feedback modeling

From test suite: comparison w/ analytical result (J. Comp. Phys, 211,1, 2005)

Synergia used in international space charge benchmark effort

lead by I.Hofmann (PAC'05)

Unique capabilities for synchrotrons, boosters, and storage rings

Longitudinal phase space shows halo & space-charge “drag” during bunch merge

Page 23: SciDAC - accelconf.web.cern.ch

● Self-consistent 3D space-charge model

● 2nd order maps (arbitrary order possible)

● Use 33x33x257 grid and ~ 5,000,000 particles

● Multi-turn injection

Synergia simulation of the Fermilab booster: Multi-bunch modeling in 3D

Merging of 5 linac microbunches in the FNAL booster (J. Amundson, P.

Spentzouris)

•Machine ramping with position feedback•6-D phase space beam matching

Page 24: SciDAC - accelconf.web.cern.ch

Booster simulations compared with experiment

3D Booster simulation including injection, rf ramping, etc.

Comparison with experimental data

* dataסּ Synergia (envelope FFT)

Space charge tune shift using a coasting beam and scanning the

half integer resonance

Page 25: SciDAC - accelconf.web.cern.ch

Qualitative beam loss prediction

Booster close to sum resonance

Page 26: SciDAC - accelconf.web.cern.ch

Significant progress in laser/plasma based accelerator concepts

• 3 Key breakthroughs

Observation of low-energy spread bunches from LWFA

Production of 1 GeV beam from LWFA

Doubling of a 28.5 GeV beam in a PWFA

The physics behind these breakthroughs has been understood (in some cases predicted)

through large-scale PIC codes including SciDAC codes (OSIRIS, VORPAL,QuickPIC)

Page 27: SciDAC - accelconf.web.cern.ch

Plasm a-Based Acceleration

• Lim ited by peak power and breakdow n

• 20-100 MeV/m

• No breakdown lim it

• 10-100 GeV/m

Conventional Accelerators Plasm a

• Plasma Wake Field Accelerator (PWFA)A high energy electron bunch

• Laser Wake Field Accelerator (LWFA) A single short-pulse of photons

Page 28: SciDAC - accelconf.web.cern.ch

Breakthrough: Production of low energy spread beams from a LWFA

SciDAC codes used to successfully model LWFA experiments (VORPAL simulation, J. Cary/TechX)

SciDAC codes used to explore and discover paths to 1 GeV and beyond (W. Mori, OSIRIS simulation)

Page 29: SciDAC - accelconf.web.cern.ch

LWFA: prediction, experiment, verification

OSIRISOSIRISOSIRIS simulation predicting short electron bunch if

acceleration path matched to de-phasing length.

Experimental measurement and VORPAL simulation

Page 30: SciDAC - accelconf.web.cern.ch

FFTB

Breakthrough: Energy ~doubling in a PWFA:Modeling afterburner-related experiments (E167) w/ QuickPIC

Cerenkov image (top, experiment) and pz-x phase space (bottom, simulation)

In the most recent energy doubling experiment, the observed results (e.g. the final energy spectrum) agreed very well with QuickPIC predictions.

The simulations played an important role in elucidate the underlying physical mechanism (e.g. beam head erosion) for experimental observations.

Page 31: SciDAC - accelconf.web.cern.ch

• Energy of 500GeV beam is doubled in only 25 meters

• Despite some hosing wake structure remains stable

• 1TeV energy with 5% energy spread

Beam driver

Energy of drive and trailing beam afters = 25 m

Accelerating field remains stable

Virtual demonstration of a TeV afterburner

Page 32: SciDAC - accelconf.web.cern.ch

Future: SciDAC2 impact of petascale computing

• Under SciDAC2 we will take advantage of the extraordinary opportunities of petascale computing

• Develop a comprehensive accelerator modeling framework capable of multi-physics simulation necessary for end-to-end modeling

• Develop common interfaces for code interoperation

Using SciDAC2 applied math and computer

science tools, integrate algorithm optimization into

the simulation environment, at the user

level

Page 33: SciDAC - accelconf.web.cern.ch

Explore new synergies

Utilize fully self consistent EM+particle PIC codes (VORPAL) for EM design:self-consistent propagation in loaded cavities• secondary emission from cavities• feedback systems• beam diagnostics• mode and wakefield

calculations

Fully 3D Tesla cavity VORPAL simulation

Page 34: SciDAC - accelconf.web.cern.ch

WARP-POSINST hybrid code + Adaptive Mesh Refinement

Z

R

concentrates resolution only where it is needed

Speed-up x10-104

beam

quad

e- motion in a quad

New e- moverAllows large time step greater than cyclotron period

Speed-up x10-100

HCX dedicated setup for gas/electron effects studies

Page 35: SciDAC - accelconf.web.cern.ch

6 MHz signal in clearing electrode: simulation & expt

WARP-3DT = 4.65µs

Oscillations

Beam ions hit end

plate

(a) (b) (c)

e-

0V 0V 0V/+9kV 0V

Q4Q3Q2Q1

200mA K+

200mA K+

Electrons

Electrons bunching

0. 2. time (µs) 6.

��(c)

Simulation Experiment

0.

-20.

-40.

I (m

A)

Illustrates the benefits of collaboration in computational accelerator physics: An OFES code (WARP), combined w/ a OHEP code (POSINST)

into a unique parallel capability that will serve both communities.

M. Furman, J.-L. Vay (LBNL); A. Friedman, R. Cowan (LLNL); P. Stoltz (Tech-X); P. Colella (LBNL)

• Good test of secondary module - secondary electron emission turned off:

• run time ~3 days, - without new electron mover and MR, run time would be ~1-2 months!

• Good test of secondary module - secondary electron emission turned off:

• run time ~3 days, - without new electron mover and MR, run time would be ~1-2 months!

Simulation Experiment

(c)0. 2. time (µs) 6.

I (m

A)

0.

-20.

-40.

Excellent agreement withSecondary emission model turned on

Page 36: SciDAC - accelconf.web.cern.ch

Application Focus

• Accelerator design and optimization—HEP: ILC, LHC, Tevatron, PEP-II, proton

drivers

—NP: RHIC, RIA, CEBAF—BES: SNS, 4th generation light sources

emphasis on problems which require large scale (parallel) computingwork closely with machine designers and

operators to apply the tools

Page 37: SciDAC - accelconf.web.cern.ch

Application Focus

• Push the energy frontier in advanced accelerators—Explore TeV scale afterburners with high

fidelity—Explore 10 GeV+ LWFA systems

• Provide the tools for near real-time feedback to accelerator control rooms and beam-based experiments

SciDAC2 collaboration: J. Cary (Tech-X), K. Ko(SLAC), W. Mori (UCLA), R. Ryne (LBNL), P.

Spentzouris (FNAL)