View
35
Download
0
Category
Preview:
DESCRIPTION
Prospects on Texas High Energy Physics Network Needs. Jae Yu Univ. of Texas, Arlington. LEARN Strategy Meeting University of Texas at El Paso Dec. 9, 2004. Outline. High Energy Physics The challenges HEP in Texas Network Needs for HEP Conclusions. High Energy Physics. - PowerPoint PPT Presentation
Citation preview
HEP Prospects, J. Yu LEARN Strategy Meeting
Prospects on Texas High Energy Physics Network Needs
LEARN Strategy MeetingUniversity of Texas at El
PasoDec. 9, 2004
Jae YuUniv. of Texas, Arlington
Dec. 9, 2004 2HEP Prospects, J. Yu LEARN Strategy Meeting
Outline
• High Energy Physics• The challenges• HEP in Texas• Network Needs for HEP• Conclusions
Dec. 9, 2004 3HEP Prospects, J. Yu LEARN Strategy Meeting
High Energy Physics• Definition: A field of Physics pursues for fundamental
constituents of matter and basic principles of interactions between them How is universe created, and how does it work?
• Use large particle accelerators • Use large particle detectors
Dec. 9, 2004 4HEP Prospects, J. Yu LEARN Strategy Meeting
Fermilab Tevatron• World’s Highest Energy proton-anti-proton collider
– Ecm=1.96 TeV (=6.3x10-7J/p 13M Joules on 10-4m2)Equivalent to the kinetic energy of a 20t truck at a speed 80 mi/hr
Chicago
Tevatron p
p CDF
DØ
Dec. 9, 2004 5HEP Prospects, J. Yu LEARN Strategy Meeting
High Energy Physics• Definition: A field of Physics pursues for fundamental
constituents of matter and basic principles of interactions between them How is universe created, and how does it work?
• Use large particle accelerators • Use large particle detectors• Large, distributed collaborations
– ~600/experiment for currently operating experiments– ~2000/experiment for future experiments – WWW grew out of HEP to expedite communication
between collaborators
Dec. 9, 2004 6HEP Prospects, J. Yu LEARN Strategy Meeting
~700 Collaborators~80 Institutions18 Countries
Typical HEP Collaboration at Present
Dec. 9, 2004 7HEP Prospects, J. Yu LEARN Strategy Meeting
First Beams: Summer 2007Physics Runs: from Fall 2007
TOTEM
LHCb: B-physics
ALICE : HI
pp s =14 TeV L=1034 cm-2 s-1
27 km Tunnel in Switzerland & France
Large Hadron Collider (LHC) CERN, Geneva: 2007 Start
Large Hadron Collider (LHC) CERN, Geneva: 2007 Start
CMS
Atlas
5000+ Physicists 250+ Institutes 60+ Countries
H. Newman
Dec. 9, 2004 8HEP Prospects, J. Yu LEARN Strategy Meeting
High Energy Physics• Definition: A field of Physics pursues for fundamental constituents of
matter and basic principles of interactions between them How is universe created, and how does it work?
• Use large particle accelerators • Use large particle detectors• Large, distributed collaborations
– WWW grew out of HEP to expedite communication between collaborators – ~600/experiment for currently operating experiments– ~2000/experiment for future experiments
• Multi-peta bytes of data– Present experiments: ~10PB by 2008– Future experiments: Tens of PB by 2008 and ~ Exa-bytes by 2015
• Shares many of these challenges with other fields• Grid computing is adopted to provide a solution for these challenges
Dec. 9, 2004 9HEP Prospects, J. Yu LEARN Strategy Meeting
LHC Data Grid Hierarchy
Tier 1
Tier2 Center
Online System
CERN Center PBs of Disk;
Tape Robot
FNAL CenterIN2P3 Center INFN Center RAL Center
InstituteInstituteInstituteInstitute
Workstations
~100-1500 MBytes/sec
~10 Gbps
1 to 10 Gbps
Tens of Petabytes by 2007-8.An Exabyte ~5-7 Years later.Physics data cache
~PByte/sec
10 - 40 Gbps
Tier2 CenterTier2 CenterTier2 Center
~1-10 Gbps
Tier 0 +1
Tier 3
Tier 4
Tier2 Center Tier 2
Experiment
CERN/Outside Resource Ratio ~1:2Tier0/( Tier1)/( Tier2) ~1:1:1
H. Newman
HEP Prospects, J. Yu LEARN Strategy Meeting
• 30 sites, multi-VO• shared resources• ~3000 CPUs
Jan. 2004
Sep. 2004
A.SillGriPhyn Grid3 OSG0
ot
n
Grid3 Deployment Map• Over 100 users• ~100% utilization• Continued increase
Dec. 9, 2004 11HEP Prospects, J. Yu LEARN Strategy Meeting
High Energy Physics in Texas• 12 Universities in Texas are involved in HEP
– UTA, UT, UH, Rice, TTU, TAMU, UTB, UTEP, SMU, UTD, UTPA, and PVAMU
• Many different research facilities used– US: Fermi National Accelerator Laboratory, Jefferson Lab,
Brookhaven National Lab, SLAC and Cornell– Europe: CERN in Switzerland and DESY in Germany– Asia: KEK in Japan and BES in China – Natural sources of particle beams
• Sizable community, variety of experiments and needs
Dec. 9, 2004 12HEP Prospects, J. Yu LEARN Strategy Meeting
Universities in Texas w/ HEP Active Program
Dec. 9, 2004 13HEP Prospects, J. Yu LEARN Strategy Meeting
HEP Experiment Involvements and Activities• Presently operating experiments
– DØ: UTA, Rice, SMU• UTA has the only DØ Regional Analysis Center• JY at UTA playing a leadership role in DØ grid computing
– CDF: TTU, TAMU, Baylor• A. Sill from TTU is the CDF grid computing coordinator
– Babar: UH, UTD– MINOS: UT
• LHC Experiments– ATLAS: UTA, SMU, UTD, UTPA
• K. De at UTA a grid computing leader• UTA a candidate for a Tier 2 center
– CMS: RICE, TTU– ALICE: UH
• L. Pinsky at UH the computing coordinator of ALICE• UH is competing for Tier 1 center for ALICE
– LHCB: UT• Other Current experiments
– STAR (Rice, TAMU)– Belle (UT)
• Beyond the next decade: Linear Collider• Texas HEP grid: THEGrid as part of HiPCAT
Dec. 9, 2004 14HEP Prospects, J. Yu LEARN Strategy Meeting
UTA – RAC (DPCC)•100 P4 Xeon 2.6GHz CPU = 260 GHz•64TB of Disk space
•84 P4 Xeon 2.4GHz CPU = 202 GHz•7.5TB of Disk space
•Total CPU: 300k SI2000•Total disk: 73TB•Total Memory: 168Gbyte•Network bandwidth: 68Gb/sec
Dec. 9, 2004 15HEP Prospects, J. Yu LEARN Strategy Meeting
Network Bandwidth Usage at UTA
DPCC online
DØ and ATLAS Production
Dec. 9, 2004 16HEP Prospects, J. Yu LEARN Strategy Meeting
Network Needs for HEP• For current experiments
– DØ Regional Center Resources• 200k SI2000 (UTA) 125Mbit/s peak and 30Mbits/sec average
– Other experiments 155 Mbits/s average • For future experiments
– Anticipated needs to support all experiments on 2008 Optimal average bandwidth 622Mbit/s
– Additional needs if large hubs get located in Texas• Anticipated future ATLAS tier 2 resources
– 2005: 300k SI2000– 2008 and future: 3000k SI2000
• ALICE Tier one will add just as much as ATLAS does
– Optimal average: 1 – 2 Gbit/s
Dec. 9, 2004 17HEP Prospects, J. Yu LEARN Strategy Meeting
Conclusions• Texas HEP community plays leadership roles in virtually all
present and future experiments• Grid computing from the necessity promotes
interdisciplinary research activities – Good high level work force training– Attract external funds
• High bandwidth a key infrastructure for maintaining leadership in HEP and computing
• LEARN’s planned network and its expeditious implementation to support HEP activities critical in this endeavor
Recommended