Beam dynamics calculation

Size: px
Start display at page:

Download "Beam dynamics calculation"

Transcription

1 September 6 Beam dynamics calculation S.B. Vorozhtsov, Е.Е. Perepelkin and V.L. Smirnov Dubna, JINR

2 Outline Problem formulation Numerical methods OpenMP and CUDA realization Results CBDA: Cyclotron Beam Dynamic Analysis code

3 Computer model of the cyclotron Injection line Inflector Dee ESD Magnet sectors

4 Inflector

5 Central region Courtesy A.S. Vorozhtsov

6 Regions of the field maps Inflector Electric field Axial channel Magnet field G1 Magnet field

7 Motion equation r r, p, t, i= 1K N ( ) i i d r r p i = F i, i = 1K N dt d r r r r r r r r mi ( γ ivi ) = qi Eext ( i, t) + Es ( i, t) + vi, Bext ( i ) dt 2 vi γi = 1 1 βi, βi = c r r( 0) r r( 0) r r( 0) r r( 0) i = i, vi = vi, или i =, t ti t ti t 0 i vi = v = = = t= 0 i r 1 i N, i V ( )

8 Space charge effect r ρ r r 1 r dive =, rotb = μ J + E s s 0 s 2 s ε 0 c t r r r 1 rotes = Bs, divbs = 0, = c t με PIC метоdays r E s = ϕ ( p) ρ Δ ϕ ( p) =, p Ω ε 0 ϕ ϕ = ϕd, = ψ N, Γ Γ D Γ N =Γ D n ΓN PP метоdays r N r 1 q j r r E ( ) = r r 3 ( ), i= 1K N s i i j 4πε 0 j i ri rj r 1 q r r r r E j i =, < R j ( ) ( ) s 3 i j i j 4πε 0 R

9 Space charge effect ( p) ρ Δ ϕ ( p) =, p Ω ε 0 ϕ Γ = 0 ρ z y x ( nmk) ρ( xi yj zs) ϕ ( i, j, s) ρ x y z obtain from the particle distribution N 1 N 1 N 1 8 πni πmj πks,, =,, sin sin sin NNN x y z s= 1 j= 1 i= 1 N x N y Nz πn πm πk = + + L x L y Lz ( nmk,, ) ρ( nmk,, ) FFT use for Furier set N 1 z 1 N y Nx 1 πni πmj πks ϕ( xi, yj, zs) = ϕ( n, m, k) sin sin sin k= 1 m= 1 n= 1 N x N y Nz 1

10 Bunch area Lz Lx Lz Ly Lx Lx Ly Lz Ly Mesh Nx Ny Nz Lx Ly Lz Mesh steps hx =, hy =, hz = Nx Ny Nz

11 Charge density calculation Cell 7 Cell 8 Cell 6 Cell 3 Node Cell 5 Cell 2 Cell 1

12 Particle losses t n+1 D B A C If point D belongs the ABC triangle then Cross condition S + S + S = S ΔADC ΔADB ΔCDB ΔABC t n S + S + S < S + ε ΔADC ΔADB ΔCDB ΔABC Δ where ε Δ maximal deviation from surface

13 Central region optimization «Ground» φ RF = 15 φ RF = 13 without posts F = ZU RF -W GAP Dee φ RF = 28 φ RF = 10 with posts

14 Choice optimal configuration S0 S1 S2 S3 S4

15 Accelerating field distribution

16 Very time consuming problem About 5 different variants - minimum Many ions species - accelerated Very complicated structure Multi macro particle simulation for SC dominated beams One run requires ~ several days of computer time

17 Numerical methods Runge-Kutta 4 th order ( motion equations ) FFT for Poisson equation ( space charge effect ) Beam losses ( «ray tracing» algorithm )

18 Open Multi-Processing OpenMP

19 Spiral inflector

20 Beam parameters at the inflector entrance

21 Beam parameters at the inflector exit Blue points PIC by FFT ( mesh: 2 5 x 2 5 x 2 5 ) Red points PP

22 Calculation time Method Without OpenMP With OpenMP Computer platform PP 4 h. 53 min. 2 h. 34 min. 4 h. 38min 1 h. 25 min. AMD Turion 64 2, 1.60 GHz Intel Core2 Quad 2.4 GHz PIC ~11 min. ~6 min. 2 5 x 2 5 x min. ~2 min. AMD Turion 64 2, 1.60 GHz Intel Core2 Quad 2.4 GHz 10,000 particles No geometry losses

23 CUDA Compute Unified Device Architecture

24 CUDA kernel-functions Track ( field maps, particle coordinate and velocity ) Losses (setup geometry, particle coordinate ) Rho ( particle coordinate ) FFT ( charge density function ) PoissonSolver ( Furies coefficients ) E_SC ( electric field potentials )

25 global void Track ( ) Function with many parameters. Use variable type constant : device constant float d_float[200]; device constant int d_int[80]; Particle number corresponds: int n = threadidx.x+blockidx.x*blockdim.x; Number of if, goto, for should be dicreased Linear interpolation field maps texture < float, 2, cudareadmodeelementtype > Ex_TexRef; Ex_TexRef.filterMode = cudafiltermodelinear; ( ) ( ) ( ) P = 1 α P1+ αp2 1 β + 1 α P3+ αp4 β 0 α, β 1 v α 1-α P1 P3 P2 P4 β 1-β u

26 global void Losses ( ) Geometry structure consist of triangles. Block threads copies the triangle nodes from global to shared memory. Threads synchronize after copy operation syncthreads() Particle number corresponds to thread number: int n = threadidx.x+blockidx.x*blockdim.x; Check particles and triangles match

27 global void Rho Calculate charge impact in nodes of the mesh from particle with number n = threadidx.x + blockidx.x*blockdim.x Particle gives impact to 8 nodes. Therefore, each thread store 16 values: 8 numbers of node and 8 values of charge density. For each node summarize all impacts of charge density. Cell 7 Cell 8 Cell 6 Cell 3 Node Cell 5 Cell 2 Cell 1

28 global FFT ( ) Used real FFT for sin(πn/n) basis function 3D transform consist of three 1D FFT for each axies: X, Y, Z int n = threadidx.x+blockidx.x*blockdim.x; k=(int)(n/(ny+1)); j=n-k*(ny+1); m=j*(nx+1)+k*(nx+1)*(ny+1); FFT_X[i+1]=Rho[i+m]; n = j + k*(ny+1) NY NZ Rho function data ( 3D massive )

29 global PoissonSolver ( ) Thread number int n = threadidx.x+blockidx.x*blockdim.x; Each thread calculate Furier coefficient PhiF of potential function Phi PhiF ind(i,j,k) = -RhoF ind(i,j,k) / ( kx i2 + ky j2 + kz k2 ) in the node: ind(i,j,k)=i+j*(nx+1)+k*(nx+1)*(ny+1), where k=(int)(n/(nx+1)*(ny+1)); j=(int)(n-k*(nx+1)*(ny+1))/(nx+1); i=n-j*(nx+1)-k*(nx+1)*(ny+1); RhoF Furier coefficients of charge density function Rho.

30 global E_SC ( ) Electric field calculation at mesh node int n = threadidx.x+blockidx.x*blockdim.x+st_ind φ n + ( NX + 1 )( NY + 1 ) φ φ n n + ( NX + 1 ) φ n - 1 φ n - ( NX + 1 ) E E E x y z ϕ = ϕ = ϕ = ϕ n+ 1 n 1 2h x ϕ ( 1) ( 1) n+ Nx+ n Nx+ 2h y ϕ ( 1)( 1) ( 1)( 1) n+ Nx+ Ny+ n Nx+ Ny+ 2h z φ n + 1 φ n - ( NX + 1 )( NY + 1 )

31 Particle losses

32 Bunch acceleration

33 GeForce 8800GTX CUDA kernel-function* Time, [ms] CPU ** GPU Rate, [x] Track Losses Rho Poisson/FFT E_SC Total *Mesh size: 2 5 x 2 5 x 2 5. Number of particles: 100,000 triangles: 2054 **CPU 2.4 GHz

34 GeForce 8800GTX Number of particles Time calculation CPU * GPU Rate, [x] 1,000 3 min. 19 sec. 12 sec , min. 14 sec. 42 sec ,000 5h.41 min. 6 min. 56 1,000,000 2 days 8 h. 53 min. 1h. 60 *CPU 2.4 GHz

35 Tesla C1060 Number of particles Time calculation CPU GPU 2.5GHz C 1060 Rate, [x] 1,000 3 min. 12 sec. 11 sec , min. 24 sec. 27 sec ,000 5 h. 14 min. 31 sec. 3 min. 34 sec. 88 1,000,000 2 days 4 h. 25 min. 34 min. 29 sec. 91 without space charge effect

36 Tesla C1060 Number of particles Time calculation CPU GPU 2.5 GHz C 1060 Rate, [x] 10, min. 36 sec. 44 sec ,000 5 h. 28 min. 12 sec. 5 min. 4 sec. 65 1,000,000 2 days 8 h. 27 min. 50 min. 17 sec. 67 with space charge effect

37 Space charge effect I ~ 0 Losses 24% I = 4 mа Losses 94%

38 Conclusions Inexpensive technology. Increased performance times gives a chance to perform the whole cyclotron computer modeling. Very careful programming is required that is the price for the obtained high performance of calculations. Application to other complicated simulation, related to the accelerator physics (beam halo etc), is possible.

39 Education and Research Center «Applied Parallel Compute»

Population annealing study of the frustrated Ising antiferromagnet on the stacked triangular lattice

Population annealing study of the frustrated Ising antiferromagnet on the stacked triangular lattice Population annealing study of the frustrated Ising antiferromagnet on the stacked triangular lattice Michal Borovský Department of Theoretical Physics and Astrophysics, University of P. J. Šafárik in Košice,

More information

Progress in April December 2007 Schedule in Jan. March 2008

Progress in April December 2007 Schedule in Jan. March 2008 A Report to the Advisory Committee of CNS The Accelerator Group Outline Upgrade of AVF Cyclotron Progress in April December 2007 Schedule in Jan. March 2008 Shigeru Kubono, Yukimitsu Ohshiro, Shin-ichi

More information

Multicore Parallelization of Determinant Quantum Monte Carlo Simulations

Multicore Parallelization of Determinant Quantum Monte Carlo Simulations Multicore Parallelization of Determinant Quantum Monte Carlo Simulations Andrés Tomás, Che-Rung Lee, Zhaojun Bai, Richard Scalettar UC Davis SIAM Conference on Computation Science & Engineering Reno, March

More information

Parallel programming practices for the solution of Sparse Linear Systems (motivated by computational physics and graphics)

Parallel programming practices for the solution of Sparse Linear Systems (motivated by computational physics and graphics) Parallel programming practices for the solution of Sparse Linear Systems (motivated by computational physics and graphics) Eftychios Sifakis CS758 Guest Lecture - 19 Sept 2012 Introduction Linear systems

More information

arxiv: v1 [hep-lat] 7 Oct 2010

arxiv: v1 [hep-lat] 7 Oct 2010 arxiv:.486v [hep-lat] 7 Oct 2 Nuno Cardoso CFTP, Instituto Superior Técnico E-mail: nunocardoso@cftp.ist.utl.pt Pedro Bicudo CFTP, Instituto Superior Técnico E-mail: bicudo@ist.utl.pt We discuss the CUDA

More information

SPARSE SOLVERS POISSON EQUATION. Margreet Nool. November 9, 2015 FOR THE. CWI, Multiscale Dynamics

SPARSE SOLVERS POISSON EQUATION. Margreet Nool. November 9, 2015 FOR THE. CWI, Multiscale Dynamics SPARSE SOLVERS FOR THE POISSON EQUATION Margreet Nool CWI, Multiscale Dynamics November 9, 2015 OUTLINE OF THIS TALK 1 FISHPACK, LAPACK, PARDISO 2 SYSTEM OVERVIEW OF CARTESIUS 3 POISSON EQUATION 4 SOLVERS

More information

Acceleration of Deterministic Boltzmann Solver with Graphics Processing Units

Acceleration of Deterministic Boltzmann Solver with Graphics Processing Units Acceleration of Deterministic Boltzmann Solver with Graphics Processing Units V.V.Aristov a, A.A.Frolova a, S.A.Zabelok a, V.I.Kolobov b and R.R.Arslanbekov b a Dorodnicn Computing Centre of the Russian

More information

Research of the new Intel Xeon Phi architecture for solving a wide range of scientific problems at JINR

Research of the new Intel Xeon Phi architecture for solving a wide range of scientific problems at JINR Research of the new Intel Xeon Phi architecture for solving a wide range of scientific problems at JINR Podgainy D.V., Streltsova O.I., Zuev M.I. on behalf of Heterogeneous Computations team HybriLIT LIT,

More information

A CUDA Solver for Helmholtz Equation

A CUDA Solver for Helmholtz Equation Journal of Computational Information Systems 11: 24 (2015) 7805 7812 Available at http://www.jofcis.com A CUDA Solver for Helmholtz Equation Mingming REN 1,2,, Xiaoguang LIU 1,2, Gang WANG 1,2 1 College

More information

Explore Computational Power of GPU in Electromagnetics and Micromagnetics

Explore Computational Power of GPU in Electromagnetics and Micromagnetics Explore Computational Power of GPU in Electromagnetics and Micromagnetics Presenter: Sidi Fu, PhD candidate, UC San Diego Advisor: Prof. Vitaliy Lomakin Center of Magnetic Recording Research, Department

More information

Small Isochronous Ring (SIR) project at NSCL, MSU. Eduard Pozdeyev NSCL, Michigan Sate University

Small Isochronous Ring (SIR) project at NSCL, MSU. Eduard Pozdeyev NSCL, Michigan Sate University Small Isochronous Ring (SIR) project at NSCL, MSU Eduard Pozdeyev NSCL, Michigan Sate University Talk Outline Isochronous regime in accelerators, application to Isochronous Cyclotrons Space charge effects

More information

Magnetic field simulation in the central region of the VINCY Cyclotron

Magnetic field simulation in the central region of the VINCY Cyclotron NUKLEONIKA 2003;48(Supplement 2):S39 S44 PROCEEDINGS Magnetic field simulation in the central region of the VINCY Cyclotron Sergei B. Vorojtsov, Aleksei S. Vorozhtsov, Nebojsa Nešković, Jasna Ristić-Ðurović,

More information

Using a CUDA-Accelerated PGAS Model on a GPU Cluster for Bioinformatics

Using a CUDA-Accelerated PGAS Model on a GPU Cluster for Bioinformatics Using a CUDA-Accelerated PGAS Model on a GPU Cluster for Bioinformatics Jorge González-Domínguez Parallel and Distributed Architectures Group Johannes Gutenberg University of Mainz, Germany j.gonzalez@uni-mainz.de

More information

Two case studies of Monte Carlo simulation on GPU

Two case studies of Monte Carlo simulation on GPU Two case studies of Monte Carlo simulation on GPU National Institute for Computational Sciences University of Tennessee Seminar series on HPC, Feb. 27, 2014 Outline 1 Introduction 2 Discrete energy lattice

More information

Multi-GPU Simulations of the Infinite Universe

Multi-GPU Simulations of the Infinite Universe () Multi-GPU of the Infinite with with G. Rácz, I. Szapudi & L. Dobos Physics of Complex Systems Department Eötvös Loránd University, Budapest June 22, 2018, Budapest, Hungary Outline 1 () 2 () Concordance

More information

Scalable and Power-Efficient Data Mining Kernels

Scalable and Power-Efficient Data Mining Kernels Scalable and Power-Efficient Data Mining Kernels Alok Choudhary, John G. Searle Professor Dept. of Electrical Engineering and Computer Science and Professor, Kellogg School of Management Director of the

More information

Accelerating Model Reduction of Large Linear Systems with Graphics Processors

Accelerating Model Reduction of Large Linear Systems with Graphics Processors Accelerating Model Reduction of Large Linear Systems with Graphics Processors P. Benner 1, P. Ezzatti 2, D. Kressner 3, E.S. Quintana-Ortí 4, Alfredo Remón 4 1 Max-Plank-Institute for Dynamics of Complex

More information

Dynamic Scheduling for Work Agglomeration on Heterogeneous Clusters

Dynamic Scheduling for Work Agglomeration on Heterogeneous Clusters Dynamic Scheduling for Work Agglomeration on Heterogeneous Clusters Jonathan Lifflander, G. Carl Evans, Anshu Arya, Laxmikant Kale University of Illinois Urbana-Champaign May 25, 2012 Work is overdecomposed

More information

Welcome to MCS 572. content and organization expectations of the course. definition and classification

Welcome to MCS 572. content and organization expectations of the course. definition and classification Welcome to MCS 572 1 About the Course content and organization expectations of the course 2 Supercomputing definition and classification 3 Measuring Performance speedup and efficiency Amdahl s Law Gustafson

More information

sri 2D Implicit Charge- and Energy- Conserving Particle-in-cell Application Using CUDA Christopher Leibs Karthik Murthy

sri 2D Implicit Charge- and Energy- Conserving Particle-in-cell Application Using CUDA Christopher Leibs Karthik Murthy 2D Implicit Charge- and Energy- Conserving sri Particle-in-cell Application Using CUDA Christopher Leibs Karthik Murthy Mentors Dana Knoll and Allen McPherson IS&T CoDesign Summer School 2012, Los Alamos

More information

DEVELOPMENT OF JINR FLNR HEAVY-ION ACCELERATOR COMPLEX IN THE NEXT 7 YEARS

DEVELOPMENT OF JINR FLNR HEAVY-ION ACCELERATOR COMPLEX IN THE NEXT 7 YEARS Ó³ Ÿ. 2010.. 7, º 7(163).. 827Ä834 ˆ ˆŠ ˆ ˆŠ Š ˆ DEVELOPMENT OF JINR FLNR HEAVY-ION ACCELERATOR COMPLEX IN THE NEXT 7 YEARS G. Gulbekyan, B. Gikal, I. Kalagin, N. Kazarinov Joint Institute for Nuclear

More information

On Verification and Validation of Spring Fabric Model

On Verification and Validation of Spring Fabric Model On Verification and Validation of Spring Fabric Model Zheng Gao, Qiangqiang Shi, Yiyang Yang, Bernard Moore, and Xiaolin Li Department of Applied Mathematics and Statistics Stony Brook University Stony

More information

APPLICATION OF CUDA TECHNOLOGY FOR CALCULATION OF GROUND STATES OF FEW-BODY NUCLEI BY FEYNMAN'S CONTINUAL INTEGRALS METHOD

APPLICATION OF CUDA TECHNOLOGY FOR CALCULATION OF GROUND STATES OF FEW-BODY NUCLEI BY FEYNMAN'S CONTINUAL INTEGRALS METHOD APPLICATION OF CUDA TECHNOLOGY FOR CALCULATION OF GROUND STATES OF FEW-BODY NUCLEI BY FEYNMAN'S CONTINUAL INTEGRALS METHOD M.A. Naumenko, V.V. Samarin Joint Institute for Nuclear Research, Dubna, Russia

More information

Introduction to numerical computations on the GPU

Introduction to numerical computations on the GPU Introduction to numerical computations on the GPU Lucian Covaci http://lucian.covaci.org/cuda.pdf Tuesday 1 November 11 1 2 Outline: NVIDIA Tesla and Geforce video cards: architecture CUDA - C: programming

More information

Fast Space Charge Calculations with a Multigrid Poisson Solver & Applications

Fast Space Charge Calculations with a Multigrid Poisson Solver & Applications Fast Space Charge Calculations with a Multigrid Poisson Solver & Applications Gisela Pöplau Ursula van Rienen Rostock University DESY, Hamburg, April 26, 2005 Bas van der Geer Marieke de Loos Pulsar Physics

More information

TR A Comparison of the Performance of SaP::GPU and Intel s Math Kernel Library (MKL) for Solving Dense Banded Linear Systems

TR A Comparison of the Performance of SaP::GPU and Intel s Math Kernel Library (MKL) for Solving Dense Banded Linear Systems TR-0-07 A Comparison of the Performance of ::GPU and Intel s Math Kernel Library (MKL) for Solving Dense Banded Linear Systems Ang Li, Omkar Deshmukh, Radu Serban, Dan Negrut May, 0 Abstract ::GPU is a

More information

Dense Arithmetic over Finite Fields with CUMODP

Dense Arithmetic over Finite Fields with CUMODP Dense Arithmetic over Finite Fields with CUMODP Sardar Anisul Haque 1 Xin Li 2 Farnam Mansouri 1 Marc Moreno Maza 1 Wei Pan 3 Ning Xie 1 1 University of Western Ontario, Canada 2 Universidad Carlos III,

More information

Background. Another interests. Sieve method. Parallel Sieve Processing on Vector Processor and GPU. RSA Cryptography

Background. Another interests. Sieve method. Parallel Sieve Processing on Vector Processor and GPU. RSA Cryptography Background Parallel Sieve Processing on Vector Processor and GPU Yasunori Ushiro (Earth Simulator Center) Yoshinari Fukui (Earth Simulator Center) Hidehiko Hasegawa (Univ. of Tsukuba) () RSA Cryptography

More information

Parallelism of MRT Lattice Boltzmann Method based on Multi-GPUs

Parallelism of MRT Lattice Boltzmann Method based on Multi-GPUs Parallelism of MRT Lattice Boltzmann Method based on Multi-GPUs 1 School of Information Engineering, China University of Geosciences (Beijing) Beijing, 100083, China E-mail: Yaolk1119@icloud.com Ailan

More information

Antti-Pekka Hynninen, 5/10/2017, GTC2017, San Jose CA

Antti-Pekka Hynninen, 5/10/2017, GTC2017, San Jose CA S7255: CUTT: A HIGH- PERFORMANCE TENSOR TRANSPOSE LIBRARY FOR GPUS Antti-Pekka Hynninen, 5/10/2017, GTC2017, San Jose CA MOTIVATION Tensor contractions are the most computationally intensive part of quantum

More information

Faster Kinetics: Accelerate Your Finite-Rate Combustion Simulation with GPUs

Faster Kinetics: Accelerate Your Finite-Rate Combustion Simulation with GPUs Faster Kinetics: Accelerate Your Finite-Rate Combustion Simulation with GPUs Christopher P. Stone, Ph.D. Computational Science and Engineering, LLC Kyle Niemeyer, Ph.D. Oregon State University 2 Outline

More information

GPU Accelerated Markov Decision Processes in Crowd Simulation

GPU Accelerated Markov Decision Processes in Crowd Simulation GPU Accelerated Markov Decision Processes in Crowd Simulation Sergio Ruiz Computer Science Department Tecnológico de Monterrey, CCM Mexico City, México sergio.ruiz.loza@itesm.mx Benjamín Hernández National

More information

CRYPTOGRAPHIC COMPUTING

CRYPTOGRAPHIC COMPUTING CRYPTOGRAPHIC COMPUTING ON GPU Chen Mou Cheng Dept. Electrical Engineering g National Taiwan University January 16, 2009 COLLABORATORS Daniel Bernstein, UIC, USA Tien Ren Chen, Army Tanja Lange, TU Eindhoven,

More information

Tips Geared Towards R. Adam J. Suarez. Arpil 10, 2015

Tips Geared Towards R. Adam J. Suarez. Arpil 10, 2015 Tips Geared Towards R Departments of Statistics North Carolina State University Arpil 10, 2015 1 / 30 Advantages of R As an interpretive and interactive language, developing an algorithm in R can be done

More information

Accelerating linear algebra computations with hybrid GPU-multicore systems.

Accelerating linear algebra computations with hybrid GPU-multicore systems. Accelerating linear algebra computations with hybrid GPU-multicore systems. Marc Baboulin INRIA/Université Paris-Sud joint work with Jack Dongarra (University of Tennessee and Oak Ridge National Laboratory)

More information

Calculation of ground states of few-body nuclei using NVIDIA CUDA technology

Calculation of ground states of few-body nuclei using NVIDIA CUDA technology Calculation of ground states of few-body nuclei using NVIDIA CUDA technology M. A. Naumenko 1,a, V. V. Samarin 1, 1 Flerov Laboratory of Nuclear Reactions, Joint Institute for Nuclear Research, 6 Joliot-Curie

More information

A Quantum Chemistry Domain-Specific Language for Heterogeneous Clusters

A Quantum Chemistry Domain-Specific Language for Heterogeneous Clusters A Quantum Chemistry Domain-Specific Language for Heterogeneous Clusters ANTONINO TUMEO, ORESTE VILLA Collaborators: Karol Kowalski, Sriram Krishnamoorthy, Wenjing Ma, Simone Secchi May 15, 2012 1 Outline!

More information

A Massively Parallel Eigenvalue Solver for Small Matrices on Multicore and Manycore Architectures

A Massively Parallel Eigenvalue Solver for Small Matrices on Multicore and Manycore Architectures A Massively Parallel Eigenvalue Solver for Small Matrices on Multicore and Manycore Architectures Manfred Liebmann Technische Universität München Chair of Optimal Control Center for Mathematical Sciences,

More information

Direct Self-Consistent Field Computations on GPU Clusters

Direct Self-Consistent Field Computations on GPU Clusters Direct Self-Consistent Field Computations on GPU Clusters Guochun Shi, Volodymyr Kindratenko National Center for Supercomputing Applications University of Illinois at UrbanaChampaign Ivan Ufimtsev, Todd

More information

Development of Large Scale Optimization Tools for Beam Tracking Codes

Development of Large Scale Optimization Tools for Beam Tracking Codes Development of Large Scale Optimization Tools for Beam Tracking Codes 42 nd ICFA Advanced Beam Dynamics Workshop: High-Intensity, High-Brightness Hadron Beams HB-2008, August 25-29, 2008 Nashville, Tennessee,

More information

On Portability, Performance and Scalability of a MPI OpenCL Lattice Boltzmann Code

On Portability, Performance and Scalability of a MPI OpenCL Lattice Boltzmann Code On Portability, Performance and Scalability of a MPI OpenCL Lattice Boltzmann Code E Calore, S F Schifano, R Tripiccione Enrico Calore INFN Ferrara, Italy 7 th Workshop on UnConventional High Performance

More information

Open-Source Parallel FE Software : FrontISTR -- Performance Considerations about B/F (Byte per Flop) of SpMV on K-Supercomputer and GPU-Clusters --

Open-Source Parallel FE Software : FrontISTR -- Performance Considerations about B/F (Byte per Flop) of SpMV on K-Supercomputer and GPU-Clusters -- Parallel Processing for Energy Efficiency October 3, 2013 NTNU, Trondheim, Norway Open-Source Parallel FE Software : FrontISTR -- Performance Considerations about B/F (Byte per Flop) of SpMV on K-Supercomputer

More information

Multiphase Flow Simulations in Inclined Tubes with Lattice Boltzmann Method on GPU

Multiphase Flow Simulations in Inclined Tubes with Lattice Boltzmann Method on GPU Multiphase Flow Simulations in Inclined Tubes with Lattice Boltzmann Method on GPU Khramtsov D.P., Nekrasov D.A., Pokusaev B.G. Department of Thermodynamics, Thermal Engineering and Energy Saving Technologies,

More information

Prof. Brant Robertson Department of Astronomy and Astrophysics University of California, Santa

Prof. Brant Robertson Department of Astronomy and Astrophysics University of California, Santa Accelerated Astrophysics: Using NVIDIA GPUs to Simulate and Understand the Universe Prof. Brant Robertson Department of Astronomy and Astrophysics University of California, Santa Cruz brant@ucsc.edu, UC

More information

Lecture 3: Modeling Accelerators Fringe fields and Insertion devices. X. Huang USPAS, January 2015 Hampton, Virginia

Lecture 3: Modeling Accelerators Fringe fields and Insertion devices. X. Huang USPAS, January 2015 Hampton, Virginia Lecture 3: Modeling Accelerators Fringe fields and Insertion devices X. Huang USPAS, January 05 Hampton, Virginia Fringe field effects Dipole Quadrupole Outline Modeling of insertion devices Radiation

More information

Lecture 2: Modeling Accelerators Calculation of lattice functions and parameters. X. Huang USPAS, January 2015 Hampton, Virginia

Lecture 2: Modeling Accelerators Calculation of lattice functions and parameters. X. Huang USPAS, January 2015 Hampton, Virginia Lecture 2: Modeling Accelerators Calculation of lattice functions and parameters X. Huang USPAS, January 2015 Hampton, Virginia 1 Outline Closed orbit Transfer matrix, tunes, Optics functions Chromatic

More information

Available online at ScienceDirect. Procedia Engineering 61 (2013 ) 94 99

Available online at  ScienceDirect. Procedia Engineering 61 (2013 ) 94 99 Available online at www.sciencedirect.com ScienceDirect Procedia Engineering 6 (203 ) 94 99 Parallel Computational Fluid Dynamics Conference (ParCFD203) Simulations of three-dimensional cavity flows with

More information

Investigation on the spiral inflector and central region of the IsoDAR test-bench cyclotron

Investigation on the spiral inflector and central region of the IsoDAR test-bench cyclotron Investigation on the spiral inflector and central region of the IsoDAR test-bench cyclotron Grazia D Agostino, W. Kleeven, E. Forton, V. Nuttens, S. Zaremba, L. Calabretta, A. Calanna for IsoDAR / DAEδALUS

More information

Performance and Energy Analysis of the Iterative Solution of Sparse Linear Systems on Multicore and Manycore Architectures

Performance and Energy Analysis of the Iterative Solution of Sparse Linear Systems on Multicore and Manycore Architectures Performance and Energy Analysis of the Iterative Solution of Sparse Linear Systems on Multicore and Manycore Architectures José I. Aliaga Performance and Energy Analysis of the Iterative Solution of Sparse

More information

Cyclotron Motion. We can also work-out the frequency of the cyclotron motion. f cyc =

Cyclotron Motion. We can also work-out the frequency of the cyclotron motion. f cyc = Cyclotron Motion We can also work-out the frequency of the cyclotron motion f cyc = qb 2πm Neil Alberding (SFU Physics) Physics 121: Optics, Electricity & Magnetism Spring 2010 1 / 1 Cyclotron Motion We

More information

Optimization of Particle-In-Cell simulations for Vlasov-Poisson system with strong magnetic field

Optimization of Particle-In-Cell simulations for Vlasov-Poisson system with strong magnetic field Optimization of Particle-In-Cell simulations for Vlasov-Poisson system with strong magnetic field Edwin Chacon-Golcher Sever A. Hirstoaga Mathieu Lutz Abstract We study the dynamics of charged particles

More information

Hybrid CPU/GPU Acceleration of Detection of 2-SNP Epistatic Interactions in GWAS

Hybrid CPU/GPU Acceleration of Detection of 2-SNP Epistatic Interactions in GWAS Hybrid CPU/GPU Acceleration of Detection of 2-SNP Epistatic Interactions in GWAS Jorge González-Domínguez*, Bertil Schmidt*, Jan C. Kässens**, Lars Wienbrandt** *Parallel and Distributed Architectures

More information

GANIL STATUS REPORT. B. Jacquot, F. Chautard, A.Savalle, & Ganil Staff GANIL-DSM/CEA,IN2P3/CNRS, BP 55027, Caen Cedex, France.

GANIL STATUS REPORT. B. Jacquot, F. Chautard, A.Savalle, & Ganil Staff GANIL-DSM/CEA,IN2P3/CNRS, BP 55027, Caen Cedex, France. GANIL STATUS REPORT B. Jacquot, F. Chautard, A.Savalle, & Ganil Staff GANIL-DSM/CEA,IN2P3/CNRS, BP 55027, 4076 Caen Cedex, France Abstract The GANIL-Spiral facility (Caen, France) is dedicated to the acceleration

More information

ASTRA BASED SWARM OPTIMIZATIONS OF THE BERLinPro INJECTOR

ASTRA BASED SWARM OPTIMIZATIONS OF THE BERLinPro INJECTOR ASTRA BASED SWARM OPTIMIZATIONS OF THE BERLinPro INJECTOR M. Abo-Bakr FRAAC4 Michael Abo-Bakr 2012 ICAP Warnemünde 1 Content: Introduction ERL s Introduction BERLinPro Motivation Computational Aspects

More information

Efficient Serial and Parallel Coordinate Descent Methods for Huge-Scale Convex Optimization

Efficient Serial and Parallel Coordinate Descent Methods for Huge-Scale Convex Optimization Efficient Serial and Parallel Coordinate Descent Methods for Huge-Scale Convex Optimization Martin Takáč The University of Edinburgh Based on: P. Richtárik and M. Takáč. Iteration complexity of randomized

More information

An Implementation of SPELT(31, 4, 96, 96, (32, 16, 8))

An Implementation of SPELT(31, 4, 96, 96, (32, 16, 8)) An Implementation of SPELT(31, 4, 96, 96, (32, 16, 8)) Tung Chou January 5, 2012 QUAD Stream cipher. Security relies on MQ (Multivariate Quadratics). QUAD The Provably-secure QUAD(q, n, r) Stream Cipher

More information

An FPGA Implementation of Reciprocal Sums for SPME

An FPGA Implementation of Reciprocal Sums for SPME An FPGA Implementation of Reciprocal Sums for SPME Sam Lee and Paul Chow Edward S. Rogers Sr. Department of Electrical and Computer Engineering University of Toronto Objectives Accelerate part of Molecular

More information

S0214 : GPU Based Stacking Sequence Generation For Composite Skins Using GA

S0214 : GPU Based Stacking Sequence Generation For Composite Skins Using GA S0214 : GPU Based Stacking Sequence Generation For Composite Skins Using GA Date: 16th May 2012 Wed, 3pm to 3.25pm(Adv. Session) Sathyanarayana K., Manish Banga, and Ravi Kumar G. V. V. Engineering Services,

More information

Particle-in-Cell Code BEAMPATH for Beam Dynamics Simulations in Linear Accelerators and Beamlines*

Particle-in-Cell Code BEAMPATH for Beam Dynamics Simulations in Linear Accelerators and Beamlines* SLAC-PUB-1081 5 October 004 Particle-in-Cell Code BEAMPATH for Beam Dynamics Simulations in Linear Accelerators and Beamlines* Yuri K. Batygin Stanford Linear Accelerator Center, Stanford University, Stanford,

More information

Solving PDEs with CUDA Jonathan Cohen

Solving PDEs with CUDA Jonathan Cohen Solving PDEs with CUDA Jonathan Cohen jocohen@nvidia.com NVIDIA Research PDEs (Partial Differential Equations) Big topic Some common strategies Focus on one type of PDE in this talk Poisson Equation Linear

More information

LONGITUDINAL PARTICLE TRACKING CODE FOR A HIGH INTENSITY PROTON SYNCHROTRON

LONGITUDINAL PARTICLE TRACKING CODE FOR A HIGH INTENSITY PROTON SYNCHROTRON Proceedings of HB6, Malmö, Sweden Pre-Release Snapshot 8-July-6 :3 UTC MOPR LONGITUDINAL PARTICLE TRACKING CODE FOR A HIGH INTENSITY PROTON SYNCHROTRON M. Yamamoto, Japan Atomic Energy Agency, Tokai, Ibaraki

More information

Edwin Chacon-Golcher 1, Sever A. Hirstoaga 2 and Mathieu Lutz 3. Introduction

Edwin Chacon-Golcher 1, Sever A. Hirstoaga 2 and Mathieu Lutz 3. Introduction ESAIM: PROCEEDINGS AND SURVEYS, March 2016, Vol. 53, p. 177-190 M. Campos Pinto and F. Charles, Editors OPTIMIZATION OF PARTICLE-IN-CELL SIMULATIONS FOR VLASOV-POISSON SYSTEM WITH STRONG MAGNETIC FIELD

More information

One decade of GPUs for cosmological simulations (in Strasbourg) : fortunes & misfortunes

One decade of GPUs for cosmological simulations (in Strasbourg) : fortunes & misfortunes One decade of GPUs for cosmological simulations (in Strasbourg) : fortunes & misfortunes Dominique Aubert with P. Ocvirk, J. Chardin, J. Lewis, N. Deparis (Strasbourg, F) & N. Gillet (SNS Pisa, It) CODA

More information

Julian Merten. GPU Computing and Alternative Architecture

Julian Merten. GPU Computing and Alternative Architecture Future Directions of Cosmological Simulations / Edinburgh 1 / 16 Julian Merten GPU Computing and Alternative Architecture Institut für Theoretische Astrophysik Zentrum für Astronomie Universität Heidelberg

More information

A CPU-GPU Hybrid Implementation and Model-Driven Scheduling of the Fast Multipole Method

A CPU-GPU Hybrid Implementation and Model-Driven Scheduling of the Fast Multipole Method A CPU-GPU Hybrid Implementation and Model-Driven Scheduling of the Fast Multipole Method Jee Choi 1, Aparna Chandramowlishwaran 3, Kamesh Madduri 4, and Richard Vuduc 2 1 ECE, Georgia Tech 2 CSE, Georgia

More information

Towards a highly-parallel PDE-Solver using Adaptive Sparse Grids on Compute Clusters

Towards a highly-parallel PDE-Solver using Adaptive Sparse Grids on Compute Clusters Towards a highly-parallel PDE-Solver using Adaptive Sparse Grids on Compute Clusters HIM - Workshop on Sparse Grids and Applications Alexander Heinecke Chair of Scientific Computing May 18 th 2011 HIM

More information

Practical Combustion Kinetics with CUDA

Practical Combustion Kinetics with CUDA Funded by: U.S. Department of Energy Vehicle Technologies Program Program Manager: Gurpreet Singh & Leo Breton Practical Combustion Kinetics with CUDA GPU Technology Conference March 20, 2015 Russell Whitesides

More information

Parallel Longest Common Subsequence using Graphics Hardware

Parallel Longest Common Subsequence using Graphics Hardware Parallel Longest Common Subsequence using Graphics Hardware John Kloetzli rian Strege Jonathan Decker Dr. Marc Olano Presented by: rian Strege 1 Overview Introduction Problem Statement ackground and Related

More information

Massively parallel semi-lagrangian solution of the 6d Vlasov-Poisson problem

Massively parallel semi-lagrangian solution of the 6d Vlasov-Poisson problem Massively parallel semi-lagrangian solution of the 6d Vlasov-Poisson problem Katharina Kormann 1 Klaus Reuter 2 Markus Rampp 2 Eric Sonnendrücker 1 1 Max Planck Institut für Plasmaphysik 2 Max Planck Computing

More information

Heating and current drive: Radio Frequency

Heating and current drive: Radio Frequency Heating and current drive: Radio Frequency Dr Ben Dudson Department of Physics, University of York Heslington, York YO10 5DD, UK 13 th February 2012 Dr Ben Dudson Magnetic Confinement Fusion (1 of 26)

More information

Full-wave Electromagnetic Field Simulations in the Lower Hybrid Range of Frequencies

Full-wave Electromagnetic Field Simulations in the Lower Hybrid Range of Frequencies Full-wave Electromagnetic Field Simulations in the Lower Hybrid Range of Frequencies P.T. Bonoli, J.C. Wright, M. Porkolab, PSFC, MIT M. Brambilla, IPP, Garching, Germany E. D Azevedo, ORNL, Oak Ridge,

More information

J.I. Aliaga 1 M. Bollhöfer 2 A.F. Martín 1 E.S. Quintana-Ortí 1. March, 2009

J.I. Aliaga 1 M. Bollhöfer 2 A.F. Martín 1 E.S. Quintana-Ortí 1. March, 2009 Parallel Preconditioning of Linear Systems based on ILUPACK for Multithreaded Architectures J.I. Aliaga M. Bollhöfer 2 A.F. Martín E.S. Quintana-Ortí Deparment of Computer Science and Engineering, Univ.

More information

Claude Tadonki. MINES ParisTech PSL Research University Centre de Recherche Informatique

Claude Tadonki. MINES ParisTech PSL Research University Centre de Recherche Informatique Claude Tadonki MINES ParisTech PSL Research University Centre de Recherche Informatique claude.tadonki@mines-paristech.fr Monthly CRI Seminar MINES ParisTech - CRI June 06, 2016, Fontainebleau (France)

More information

Parallelization of Molecular Dynamics (with focus on Gromacs) SeSE 2014 p.1/29

Parallelization of Molecular Dynamics (with focus on Gromacs) SeSE 2014 p.1/29 Parallelization of Molecular Dynamics (with focus on Gromacs) SeSE 2014 p.1/29 Outline A few words on MD applications and the GROMACS package The main work in an MD simulation Parallelization Stream computing

More information

Accelerators. W. Udo Schröder, 2004

Accelerators. W. Udo Schröder, 2004 1 Accelerators Overview Electrostatic Accelerators Cascade Van de Graaff V.d.G. Tandem generator Accelerator 2-3 stages steady (DC) beam, high quality focusing, energy, currents; but low energies Accelerators

More information

Paralleliza(on and Performance of the NIM Weather Model on CPU, GPU and MIC Architectures

Paralleliza(on and Performance of the NIM Weather Model on CPU, GPU and MIC Architectures Paralleliza(on and Performance of the NIM Weather Model on CPU, GPU and MIC Architectures Mark Gove? NOAA Earth System Research Laboratory We Need Be?er Numerical Weather Predic(on Superstorm Sandy Hurricane

More information

GPU Acceleration of BCP Procedure for SAT Algorithms

GPU Acceleration of BCP Procedure for SAT Algorithms GPU Acceleration of BCP Procedure for SAT Algorithms Hironori Fujii 1 and Noriyuki Fujimoto 1 1 Graduate School of Science Osaka Prefecture University 1-1 Gakuencho, Nakaku, Sakai, Osaka 599-8531, Japan

More information

Longitudinal Beam Dynamics

Longitudinal Beam Dynamics Longitudinal Beam Dynamics Shahin Sanaye Hajari School of Particles and Accelerators, Institute For Research in Fundamental Science (IPM), Tehran, Iran IPM Linac workshop, Bahman 28-30, 1396 Contents 1.

More information

Weak focusing I. mv r. Only on the reference orbit is zero

Weak focusing I. mv r. Only on the reference orbit is zero Weak focusing I y x F x mv r 2 evb y Only on the reference orbit is zero r R x R(1 x/ R) B y R By x By B0y x B0y 1 x B0 y x R Weak focusing (II) Field index F x mv R 2 x R 1 n Betatron frequency 2 Fx mx

More information

E. V. Samsonov, B. N. Gikal, O. N. Borisov, I. A. Ivanenko NUMERICAL SIMULATION OF ION ACCELERATION AND EXTRACTION IN CYCLOTRON DC-110

E. V. Samsonov, B. N. Gikal, O. N. Borisov, I. A. Ivanenko NUMERICAL SIMULATION OF ION ACCELERATION AND EXTRACTION IN CYCLOTRON DC-110 E9-2013-121 E. V. Samsonov, B. N. Gikal, O. N. Borisov, I. A. Ivanenko NUMERICAL SIMULATION OF ION ACCELERATION AND EXTRACTION IN CYCLOTRON DC-110 Submitted to Ó³ Ÿ ³ μ μ... E9-2013-121 ² μ ³μ ² μ Ê ±μ

More information

arxiv: v1 [physics.data-an] 19 Feb 2017

arxiv: v1 [physics.data-an] 19 Feb 2017 arxiv:1703.03284v1 [physics.data-an] 19 Feb 2017 Model-independent partial wave analysis using a massively-parallel fitting framework L Sun 1, R Aoude 2, A C dos Reis 2, M Sokoloff 3 1 School of Physics

More information

Information Sciences Institute 22 June 2012 Bob Lucas, Gene Wagenbreth, Dan Davis, Roger Grimes and

Information Sciences Institute 22 June 2012 Bob Lucas, Gene Wagenbreth, Dan Davis, Roger Grimes and Accelerating the Multifrontal Method Information Sciences Institute 22 June 2012 Bob Lucas, Gene Wagenbreth, Dan Davis, Roger Grimes {rflucas,genew,ddavis}@isi.edu and grimes@lstc.com 3D Finite Element

More information

Dynamics of Ions in an Electrostatic Ion Beam Trap

Dynamics of Ions in an Electrostatic Ion Beam Trap Dynamics of Ions in an Electrostatic Ion Beam Trap Daniel Zajfman Dept. of Particle Physics Weizmann Institute of Science Israel and Max-Planck Institute for Nuclear Physics Heidelberg, Germany Charles

More information

First, a look at using OpenACC on WRF subroutine advance_w dynamics routine

First, a look at using OpenACC on WRF subroutine advance_w dynamics routine First, a look at using OpenACC on WRF subroutine advance_w dynamics routine Second, an estimate of WRF multi-node performance on Cray XK6 with GPU accelerators Based on performance of WRF kernels, what

More information

Three Loose Ends: Edge Focusing; Chromaticity; Beam Rigidity.

Three Loose Ends: Edge Focusing; Chromaticity; Beam Rigidity. Linear Dynamics, Lecture 5 Three Loose Ends: Edge Focusing; Chromaticity; Beam Rigidity. Andy Wolski University of Liverpool, and the Cockcroft Institute, Daresbury, UK. November, 2012 What we Learned

More information

Multiscale Modelling, taking into account Collisions

Multiscale Modelling, taking into account Collisions Multiscale Modelling, taking into account Collisions Andreas Adelmann (Paul Scherrer Institut) October 28, 2017 Multiscale Modelling, taking into account Collisions October 28, 2017 Page 1 / 22 1 Motivation

More information

EXASCALE COMPUTING. Implementation of a 2D Electrostatic Particle in Cell algorithm in UniÞed Parallel C with dynamic load-balancing

EXASCALE COMPUTING. Implementation of a 2D Electrostatic Particle in Cell algorithm in UniÞed Parallel C with dynamic load-balancing ExaScience Lab Intel Labs Europe EXASCALE COMPUTING Implementation of a 2D Electrostatic Particle in Cell algorithm in UniÞed Parallel C with dynamic load-balancing B. Verleye P. Henry R. Wuyts G. Lapenta

More information

Basic Principle of Cyclotron

Basic Principle of Cyclotron Basic Principle of Cyclotron V. S. Pandit vspandit@gmail.com ( Outstanding Scientist (Retired), VECC, Kolkata) E.O. Lawrence originated the concept of the cyclotron accelerator in 199. It is based on a

More information

RWTH Aachen University

RWTH Aachen University IPCC @ RWTH Aachen University Optimization of multibody and long-range solvers in LAMMPS Rodrigo Canales William McDoniel Markus Höhnerbach Ahmed E. Ismail Paolo Bientinesi IPCC Showcase November 2016

More information

Research on GPU-accelerated algorithm in 3D finite difference neutron diffusion calculation method

Research on GPU-accelerated algorithm in 3D finite difference neutron diffusion calculation method NUCLEAR SCIENCE AND TECHNIQUES 25, 0501 (14) Research on GPU-accelerated algorithm in 3D finite difference neutron diffusion calculation method XU Qi ( 徐琪 ), 1, YU Gang-Lin ( 余纲林 ), 1 WANG Kan ( 王侃 ),

More information

Quantum Computer Simulation Using CUDA (Quantum Fourier Transform Algorithm)

Quantum Computer Simulation Using CUDA (Quantum Fourier Transform Algorithm) Quantum Computer Simulation Using CUDA (Quantum Fourier Transform Algorithm) Alexander Smith & Khashayar Khavari Department of Electrical and Computer Engineering University of Toronto April 15, 2009 Alexander

More information

Particle Dynamics with MBD and FEA Using CUDA

Particle Dynamics with MBD and FEA Using CUDA Particle Dynamics with MBD and FEA Using CUDA Graham Sanborn, PhD Senior Research Engineer Solver 2 (MFBD) Team FunctionBay, Inc., S. Korea Overview MFBD: Multi-Flexible-Body Dynamics Rigid & flexible

More information

Granularity effects Bas van der Geer

Granularity effects Bas van der Geer Granularity effects Bas van der Geer Marieke de Loos Steinar Wouters Edgar Vredenbregt Jom Luiten Eindhoven University The Netherlands Globular cluster Messier 2 by Hubble Space Telescope.. Located in

More information

Logo. A Massively-Parallel Multicore Acceleration of a Point Contact Solid Mechanics Simulation DRAFT

Logo. A Massively-Parallel Multicore Acceleration of a Point Contact Solid Mechanics Simulation DRAFT Paper 1 Logo Civil-Comp Press, 2017 Proceedings of the Fifth International Conference on Parallel, Distributed, Grid and Cloud Computing for Engineering, P. Iványi, B.H.V Topping and G. Várady (Editors)

More information

An Integrative Model for Parallelism

An Integrative Model for Parallelism An Integrative Model for Parallelism Victor Eijkhout ICERM workshop 2012/01/09 Introduction Formal part Examples Extension to other memory models Conclusion tw-12-exascale 2012/01/09 2 Introduction tw-12-exascale

More information

Accelerating Linear Algebra on Heterogeneous Architectures of Multicore and GPUs using MAGMA and DPLASMA and StarPU Schedulers

Accelerating Linear Algebra on Heterogeneous Architectures of Multicore and GPUs using MAGMA and DPLASMA and StarPU Schedulers UT College of Engineering Tutorial Accelerating Linear Algebra on Heterogeneous Architectures of Multicore and GPUs using MAGMA and DPLASMA and StarPU Schedulers Stan Tomov 1, George Bosilca 1, and Cédric

More information

Interpolation with Radial Basis Functions on GPGPUs using CUDA

Interpolation with Radial Basis Functions on GPGPUs using CUDA Interpolation with Radial Basis Functions on GPGPUs using CUDA Gundolf Haase in coop. with: Dirk Martin [VRV Vienna] and Günter Offner [AVL Graz] Institute for Mathematics and Scientific Computing University

More information

Morse index of figure-eight choreography for equal mass three-body problem

Morse index of figure-eight choreography for equal mass three-body problem Morse index of figure-eight choreography for equal mass three-body problem Hiroshi Fukuda College of Liberal Arts and Sciences, Kitasato University (2017.3.22) In this talk We show Morse index of Figure-eight

More information

Panorama des modèles et outils de programmation parallèle

Panorama des modèles et outils de programmation parallèle Panorama des modèles et outils de programmation parallèle Sylvain HENRY sylvain.henry@inria.fr University of Bordeaux - LaBRI - Inria - ENSEIRB April 19th, 2013 1/45 Outline Introduction Accelerators &

More information

ECE 422/522 Power System Operations & Planning/Power Systems Analysis II : 7 - Transient Stability

ECE 422/522 Power System Operations & Planning/Power Systems Analysis II : 7 - Transient Stability ECE 4/5 Power System Operations & Planning/Power Systems Analysis II : 7 - Transient Stability Spring 014 Instructor: Kai Sun 1 Transient Stability The ability of the power system to maintain synchronism

More information