Upload
archibald-tate
View
213
Download
0
Embed Size (px)
Citation preview
Midwest Tier2 Networking Status and Plans
Rob GardnerComputation and Enrico Fermi InstitutesUniversity of Chicago
USATLAS Tier1 and Tier2 Networking Planning MeetingBrookhaven LabDecember 14, 2005
12/14/05 R. Gardner - MWT2 Network Plan 2
MWT2 Personnel here today Fred Luehring: MWT2 co-PI @ IU Matt Davy: GlobalNOC @ IU Ron Rusnak: Network Services and Information Technologies (NSIT) @ UC
12/14/05 R. Gardner - MWT2 Network Plan 3
Outline Description of Midwest Tier2 (MWT2)
Network Status and Plans
12/14/05 R. Gardner - MWT2 Network Plan 4
MWT2 Description Midwest Tier2 center is a joint project between UC and IU to provide a computing resource for Monte Carlo production
facilitate physics analysis of AOD samples Project is building from iVDGL prototype Tier2 centers, which are now in place and have been active in DC2 and Rome production
This winter we will begin procurement of resources for the MWT2, adding ~32 nodes at each site and a combined 50 TB of storage, plus network upgrades
Two dedicated project-funded FTEs in place now Marty Dippel @ UC; Kristy Kallback-Rose @ IU Dan Shroeder to arrive 1/06 @ IU. Continued contributions from Andrew Zhan (GriPhyN/iVDGL)
12/14/05 R. Gardner - MWT2 Network Plan 5
The ATLAS Computing ModelJ. Shank, NSF Review 7/05
Worldwide, there are approximately 30 T2 of various sizes Approximate Overall CAPACITY in 2008:
20 MSi2k CPU 9 PB Disk
US to satisfy commitments to ATLAS: 3.3 MSi2k CPU 1.5 PB Disk
In addition, U.S. ATLAS physicists needs at our T2 will require more resources. Current estimate of our average T2 in 2008:
• 1 MSi2k CPU• 500 TB Disk• 1 MSi2k CPU• 500 TB Disk
12/14/05 R. Gardner - MWT2 Network Plan 6
Projected T2 Hardware GrowthJ. Shank, NSF Review 7/05Tier 2 Center
2005 2006 2007 2008 2009
Northeast
CPU (kSi2k)Disk (TB)
21040
350170
730370
1,090480
1,600630
Southwest
CPU (kSi2k)Disk (TB)
60060
1,000200
1,600380
1,700540
2,100700
Midwest
CPU (kSi2k)Disk (TB)
10050
240130
465260
700465
1,050790
•Assumes Moore’s law doubling of CPU and disk capacity every 3 years at constant costAssumes Moore’s law doubling of CPU and disk capacity every 3 years at constant cost•Assumes replacement of hardware every 3 yearsAssumes replacement of hardware every 3 years
12/14/05 R. Gardner - MWT2 Network Plan 7
Current Scale - Chicago 64 dual 3.0 GHz Xeon nodes, 2 GB memory Interconnected by Cisco Catalyst 3750G A storage system consists of four servers with dual 3Ware
disk controllers, providing 16 TB of attached RAID storage.
Eight front-end nodes provide grid GRAM and GridFTP servers from VDT (OSG and OSG Integration testbed gateways)
The facility also has interactive login nodes to support local batch analysis that we are providing on a best effort basis until official US ATLAS Tier2 policies are developed.
4 machine development cluster (Rocks and Condor configuration trials, OSG integration testbed)
12/14/05 R. Gardner - MWT2 Network Plan 8
Current Scale - IU The prototype Tier-2 consists of dedicated use of 64 2.4
GHz Xeon processors possible through an in-kind contribution by IU (AVIDD, NSF-MRI) to ATLAS.
The site has 1.5 TB of dedicated fiber channel disk. Recently IU provided an additional 8.0 TB of Network
Attached Storage (NAS) for ATLAS production use. Archival tape system New MWT2 equipment will be located at the ICTC
machine room on the Indianapolis campus of IU (IUPUI)
12/14/05 R. Gardner - MWT2 Network Plan 9
Network Status and Plans Overall MWT2 network architecture UC Connectivity status and plans IU Connectivity status and plans Starlight Configuration Issues IU and UC network support organizations, represented here today
12/14/05 R. Gardner - MWT2 Network Plan 11
UC Connectivity to Starlight - Status Prototype Tier2 Cisco 3750 connected via 1 Gbps path to campus border router (Cisco 6509) Once at the border, we share with the rest of UC
Campus to 710 N. Lakeshore Drive via 2 x 1 Gbps fiber provided by I-WIRE State of Illinois project: http://www.iwire.org/
At 710 NLSD DWDM output connected to MREN Force10, providing L2 Ethernet connectivity
MREN Force10 provides L3 routing 10 Gbps link between MREN and Starlight Force10, shared
12/14/05 R. Gardner - MWT2 Network Plan 12
UC-Starlight Connectivity Upgrade 10 Gbps upgrade funded by UC Teraport project (NSF-MRI)
Agreements with Starlight 9/8/05 One time cost for 10 Gbps interface at Starlight: 10K
First year of operation: 10K Components ordered from Qwest 9/15/05
DWDM 10 Gbps Lambda transceivers for UC edge routers at Starlight and at campus border: 56.1K
Delivery on these components delayed by Cisco-Qwest Current ETA: ~ month
Service brought to Research Institutes building, RI-050, where new Tier2 equipment will be housed
Tier2 to fund RI-050 connectivity w/ Cisco 6509
12/14/05 R. Gardner - MWT2 Network Plan 13
IU Connectivity to Starlight Currently there are 2x10 Gbps lambdas between IU GigaPoP and 710 NLSD On campus, 10 GigE connections from GigaPoP to ICTC, home of IU.MWT2
Connected to Starlight Force10 Current UC-IU connectivity thus goes via Abilene
IU dedicated (shared with other IU projects) in place between IUPUI and 710 NLSD Expect by Feb 06
12/14/05 R. Gardner - MWT2 Network Plan 14
Viewed from the machine room Both UC and IU are still finalizing cluster & storage design
At UC, will likely build cluster around Cisco 6509 Cisco chassis plus supervisor engine 4 port 10 Gbps interface card 1 x 48 port 1 Gbps interface card Total switching costs: ~60K, negotiating with other projects in the UC Computation Institute for sharing the costs for 10 Gbps card
Separate network for storage-compute node (backplane interference?)
At IU MWT2 infrastructure costs cover basic network connectivity
Currently, this goes via Force10, E600 Considering upgrade to E1200 Depends on final cluster and storage architecture
12/14/05 R. Gardner - MWT2 Network Plan 15
Starlight Peering Goal is to setup VLAN peering to enable 10 Gbps virtual circuits between each of the major nodes on the network: Between UC and IU hosts For either UC or IU to BNL hosts For either UC or IU to CERN hosts
For UC, setup rules on MREN router For IU, setup with dedicated IU router at Starlight
All this should be straightforward to establish for UC-IU link
Not as clear the CERN or BNL links (why we’re here!)
12/14/0512/14/05 R. Gardner - MWT2 Network PlaR. Gardner - MWT2 Network Plann
1717
NSIT – Data Networking
Networking Services and Information TechnologiesNetworking Services and Information Technologies– Academic Technologies, Administrative Systems, Data Academic Technologies, Administrative Systems, Data
Center, General Services and NetworkingCenter, General Services and Networking– ~340 individuals ~340 individuals
Data NetworkingData Networking– DirectorDirector
Installation and Repair – Assistant Director plus 5 Installation and Repair – Assistant Director plus 5 technicianstechnicians
Engineering – Manager plus 6 engineers Engineering – Manager plus 6 engineers – Equipment located in 406 closets in 128 buildings (1/4 with Equipment located in 406 closets in 128 buildings (1/4 with
UPS)UPS)– ~25 Miles of fiber between buildings~25 Miles of fiber between buildings– ~26,000 hosts~26,000 hosts– Approximately 1700 Cisco switches deployed (35xx, 2950, Approximately 1700 Cisco switches deployed (35xx, 2950,
2970)2970)– ~700 wireless APs in 97 buildings (90% of best places, 60% ~700 wireless APs in 97 buildings (90% of best places, 60%
of full coverage)of full coverage)
12/14/0512/14/05 R. Gardner - MWT2 Network PlaR. Gardner - MWT2 Network Plann
1818
NSIT – Data Networking
– 18 remote sites serviced with 2 45Mpbs and 20 1.5Mbps 18 remote sites serviced with 2 45Mpbs and 20 1.5Mbps circuits; 1 OPT-E-MAN 20Mbpscircuits; 1 OPT-E-MAN 20Mbps
– 2 ISPs on OC3 circuits2 ISPs on OC3 circuits
– Internet 2 = 3x1Gbps on I-Wire Ciena DWDM to Internet 2 = 3x1Gbps on I-Wire Ciena DWDM to MREN/StarlightMREN/Starlight Soon 10 GbpsSoon 10 Gbps
– Campus Core = 14 Cisco 6500s with Sup 720 Campus Core = 14 Cisco 6500s with Sup 720 interconnected with 1Gbpsinterconnected with 1Gbps Soon 10 GbpsSoon 10 Gbps
– Multicast SupportMulticast Support 5 large access grid facilities5 large access grid facilities
12/14/05 R. Gardner - MWT2 Network Plan 20
GlobalNOC functions24x7x365 Network Operations and Engineering ServicesReal-Time Network MonitoringDocumentation including and an extensive Network Information DatabaseProblem Tracking & ResolutionHistorical Reporting (Traffic, Availability, etc)Change ManagementSecurity Incident ResponseNetwork Design and PlanningNetwork Installation & Implementation
12/14/05 R. Gardner - MWT2 Network Plan 21
Open Questions Configuration of MWT2 connectivity with BNL via Starlight
Same, for CERN - what are the rules/policy? (for access to AOD datasets located elsewhere in ATLAS)
12/14/0512/14/05 R. Gardner - MWT2 Network PlaR. Gardner - MWT2 Network Plann
2323
I-Wire DWDM NetworkI-Wire DWDM Network
NCSA
Gleacher Center(University of Chicago)
QWest
Starlight(Northwestern
University)
University of Illinois-Chicago
Level(3)
Illinois Century Network (K-20)
University of Chicago
Illinois Institute of Technology
McLeodUSA2 DWDM Systems(660 Gb/s capacity each)
Argonne National Laboratory
Production(Abilene, ESnet, MREN,
TeraGrid, CERN, NASA…)
Experimental(CA*Net4, OmniNet, Surfnet,
Europe, Japan…)
12/14/05 R. Gardner - MWT2 Network Plan 24
Tier2 Site Architecture Provision for hosting persistent “Guest” VO services & agents
Edge Nodes
GRAM
GRIDFTP
GIP
SRM
VO Services, Agents,Proxy Caches,Catalogs
Worker nodes Worker nodes
Guest VOServices, Agents,Proxy Caches
Site Architecture
Private network Externalnetwork
Grid Gateways VO dedicatedboxes
Edge Servicesmanaged
12/14/05 R. Gardner - MWT2 Network Plan 25
Software Environment ATLAS releases provided at both MWT2 sites Distribution kit releases for production
Additional releases Local tools (UC site)
Local batch submit scripts & monitors Backups of system configuration and important service installations, and user /home implemented Have ordered a 4 TB server to provide dedicated service
12/14/05 R. Gardner - MWT2 Network Plan 26
Policy for CPU Usage Resource allocation at both UC and IU will be determined by US ATLAS policy
UC (Condor) and IU (PBS) will have different implementations, but will effectively set usage for: for production managers for software maintainers for US ATLAS users for general ATLAS for general OSG access
Set by role based authorization system (VOMS, GUMS, and local Unix accounts) Have configured UC GUMS to support US ATLAS roles
Will use UC GUMS for both MWT2 sites
Many fine graineddetails on queue and
user behavior will need to be worked
out.
Many fine graineddetails on queue and
user behavior will need to be worked
out.
12/14/05 R. Gardner - MWT2 Network Plan 27
Policy for Storage Usage User, Community and DDM (ATLAS) datasets need to be provisioned on the data servers
Storage and quota of existing 16 TB @ UC Se1: grid and system use Se2: community (local data manager) Se3: community (local data manager) Se4: user w/ quota
DDM, Panda storage at present 1 TB Expansion to 50 TB FY06 IU storage (archival)
12/14/05 R. Gardner - MWT2 Network Plan 28
Grid Environment VDT Grid clients
Provided by tier2-u2.uchicago.edu DDM clients
As part of DDM service install at UC Local specific guides
http://pgl.uchicago.edu/twiki/bin/view/Tier2/WebHome
Services deployed DDM managed agents and catalogs (still in a
long period of debugging) dCache (first deployments successful, problems
with compatibility with dCache version and FTS prevented larger scale testing)
OSG production and ITB instances
12/14/05 R. Gardner - MWT2 Network Plan 29
Service and Operations Points of contact
[email protected] (General announcements)
[email protected] (System problems)
Systems administrator response: 9-5 M-F and best effort
Non-local grid problems: US ATLAS support center @ BNL trouble ticket
IU and UC will develop MWT2 specific trouble ticketing system