UC Capabilities Supporting High-Performance Collaboration and Data-Intensive Sciences
University of California Council of Research
University of California, Irvine
October 22, 2007
Dr. Larry Smarr
Director, California Institute for Telecommunications and Information Technology
Harry E. Gruber Professor,
Dept. of Computer Science and Engineering
Jacobs School of Engineering, UCSD
Southern California Visible Satellite250,000 Evacuated
NASA Aqua MODIS Image2pm Monday PDT
Data Intensive e-Science Instruments Will Require SuperNetworks
ALMA Has a Requirement
for a 120 Gbps Data Rate per
Telescope
Access to CERN Large Hadron Collider A Terabit/s WAN by 2013!
Source: Harvey
Newman, Caltech
The Unrelenting Exponential Growth of Data Requires an Exponential Growth in Bandwidth
• “Each LHC experiment foresees a recorded raw data rate of 1 to several PetaBytes/year” – Dr. Harvey Neuman (Cal Tech), Professor of Physics
• “The VLA facility is now able to generate 700 Gbps of astronomical data and the Extended VLA will reach 3.2 Terabits per second by 2009.”– Dr. Steven Durand, National Radio Astronomy Observatory, E-VLBI Workshop,
MIT Haystack Observatory., Sep 2006
• “The Global Information Grid will need to store and access exabytes of data on a realtime basis by 2010”– Dr. Henry Dardy (DOD), Optical Fiber Conference, Los Angeles, CA USA, Mar
2006
• “US Bancorp backs up 100 TB financial data every night – now.”– David Grabski (VP Information Tech. US Bancorp), Qwest High Performance
Networking Summit, Denver, CO. USA, June 2006.
Source: Jerry Sobieski MAX / University of Maryland
fc *
Dedicated Optical Channels Makes High Performance Cyberinfrastructure Possible
(WDM)
Source: Steve Wallach, Chiaro Networks
“Lambdas”Parallel Lambdas are Driving Optical Networking
The Way Parallel Processors Drove 1990s Computing
10 Gbps per User ~ 200x Shared Internet Throughput
National LambdaRailServes the University of Virginia
UVaUCSD
“There are many potential projects that could benefit from the use of NLR,
including both high-end science projects, such as astronomy, computational biology and genomics, but also commercial applications in
the multimedia (audio and video) domain.”-- Malathi Veeraraghavan, Professor of
Electrical and Computer Engineering, UVa,PI CHEETAH Circuit Switched Testbed
The OptIPuter Project – Creating High Resolution Portals
Over Dedicated Optical Channels to Global Science Data• NSF Large Information Technology Research Proposal
– Calit2 (UCSD, UCI) and UIC Lead Campuses—Larry Smarr PI– Partnering Campuses: SDSC, USC, SDSU, NCSA, NW, TA&M, UvA,
SARA, NASA Goddard, KISTI, AIST, CRC(Canada), CICESE (Mexico)
• Engaged Industrial Partners:– IBM, Sun, Telcordia, Chiaro, Calient, Glimmerglass, Lucent
• $13.5 Million Over Five Years—Now In the Six and Final YearNIH Biomedical Informatics
Research Network NSF EarthScope and ORION
My OptIPortalTM – AffordableTermination Device for the OptIPuter Global Backplane
• 20 Dual CPU Nodes, 20 24” Monitors, ~$50,000• 1/4 Teraflop, 5 Terabyte Storage, 45 Mega Pixels--Nice PC!• Scalable Adaptive Graphics Environment ( SAGE) Jason Leigh, EVL-UIC
Source: Phil Papadopoulos SDSC, Calit2
The New Science of Metagenomics
“The emerging field of metagenomics,
where the DNA of entire communities of microbes is studied simultaneously,
presents the greatest opportunity -- perhaps since the invention of
the microscope – to revolutionize understanding of
the microbial world.” –
National Research CouncilMarch 27, 2007
NRC Report:
Metagenomic data should
be made publicly
available in international archives as rapidly as possible.
Marine Genome Sequencing Project – Measuring the Genetic Diversity of Ocean Microbes
Sorcerer II Data Will Double Number of Proteins in GenBank!
Need Ocean Data
Flat FileServerFarm
W E
B P
OR
TA
L
TraditionalUser
Response
Request
DedicatedCompute Farm
(1000s of CPUs)
TeraGrid: Cyberinfrastructure Backplane(scheduled activities, e.g. all by all comparison)
(10,000s of CPUs)
Web(other service)
Local Cluster
LocalEnvironment
DirectAccess LambdaCnxns
Data-BaseFarm
10 GigE Fabric
Calit2’s Direct Access Core Architecture Will Create Next Generation Metagenomics Server
Source: Phil Papadopoulos, SDSC, Calit2+
We
b S
erv
ice
s
Sargasso Sea Data
Sorcerer II Expedition (GOS)
JGI Community Sequencing Project
Moore Marine Microbial Project
NASA and NOAA Satellite Data
Community Microbial Metagenomics Data
“Instant” Global Microbial Metagenomics CyberCommunity
Over 1300 Registered Users From 48 Countries
USA 761United Kingdom 64Germany 54Canada 46France 44Brazil 33
CICESE
UW
JCVI
MIT
SIO UCSD
SDSU
UIC EVL
UCI
OptIPortals
OptIPortal
An Emerging High Performance Collaboratoryfor Microbial Metagenomics
UC Davis
UMich
~70 Faculty~25+ new ~700 people
Six floors225,000 sq ft$98M
Molecular MedicineGenomics & BioinformaticsPharmacologyBiomedical EngineeringEnabling Genomics FacilityImaging & Vivarium
Genome and Medical Biosciences BuildingFirst 10Gbps OptIPortal End Point at UC Davis
Building a Global Collaboratorium
Sony Digital Cinema Projector
24 Channel Digital Sound
Gigabit/sec Each Seat
Digital Cinema Standard 4k—Four Times HD!CineGrid @ iGrid2005
4K Scientific Visualization
4K Distance Learning
4K Virtual Reality
Source: Laurin Herr
4K Telepresence
4K Digital Cinema
A New Digital Medium for Art, Science,
and Collaboration
e-Science Collaboratory Without Walls Enabled by Uncompressed HD Telepresence
Photo: Harry Ammons, SDSC
John Delaney, PI LOOKING, Neptune
May 23, 2007
1500 Mbits/sec Calit2 to UW Research Channel Over NLR
Goal for SC’07iHDTV Integrated into OptIPortal
Moving from Compressed HD to Uncompressed iHDTV
Reno to UW in Seattle
Source: Michael WellingsResearch ChannelUniv. Washington
Rocks / SAGE OptIPortalsAre Being Adopted Globally
NCMIR@UCSD SIO@UCSD
UIC
Calit2@UCI
KISTI-Korea
NCSA & TRECC
Calit2@UCSD
AIST-Japan UZurich CNIC-China
NCHC-Taiwan
Osaka U-Japan
EVL’s SAGE Global Visualcasting to Europe September 2007
Image Source
OptIPuter servers at
CALIT2San Diego
Image Replication
OptIPuter SAGE-
Bridge at StarLightChicago
Image Viewing
OptIPortals at EVL
Chicago
Image Viewing
OptIPortal at SARA
Amsterdam
Image Viewing
OptIPortal at Masaryk
University Brno
Image Viewing
OptIPortal at Russian
Academy of SciencesMoscow
Oct 1
Source: Luc Renambot, EVL
Gigabit Streams
3D OptIPortals: Calit2 StarCAVE and VarrierAlpha Tests of Telepresence “Holodecks”
60 GB Texture Memory, Renders Images 3,200 Times the Speed of Single PC
Source: Tom DeFanti, Greg Dawe, Calit2Connected at 160 Gb/s
30 HD Projectors!
StarCAVE Panoramas
How Do You Get From Your Lab to the National LambdaRail?
www.ctwatch.org
“Research is being stalled by ‘information overload,’ Mr. Bement said, because data from digital instruments are piling up far faster than researchers can study. In particular, he said, campus networks need to be improved. High-speed data lines crossing the nation are the equivalent of six-lane superhighways, he said. But networks at colleges and universities are not so capable. “Those massive conduits are reduced to two-lane roads at most college and university campuses,” he said. Improving cyberinfrastructure, he said, “will transform the capabilities of campus-based scientists.”-- Arden Bement, the director of the National Science Foundation
Interconnecting Regional Optical NetworksIs Driving Campus Optical Infrastructure Deployment
http://paintsquirrel.ucs.indiana.edu/RON/fiber_map_draft.pdf
CENIC2008
1999
California (CENIC) Network Directions
• More Bandwidth to Research University Campuses – One or Two 10GE Connections to Every Campus
• More Bandwidth on the Backbone– 40Gbps Or 100Gbps
• Support for New Protocols and Features– IPv6 Multicast– Jumbo Frames: 9000 (or More) Bytes
• “Hybrid Network” Design, Incorporating Traditional Routed IP Service and the New Frame and Optical Circuit Services:– “HPRng-L3” = Routed IP Network– “HPRng-L2” = Switched Ethernet Network– “HPRng-L1” = Switched Optical Network
Source: Jim Dolgonas, CENIC
CalREN-XD
CENIC Switched Ethernet NetworkHPRng-L2 Design
Source: Jim Dolgonas, CENIC
CENIC Switched Optical NetworkHPRng-L1 design
Source: Jim Dolgonas, CENIC
Source: Jim Dolgonas, CENIC
Campus Preparations Needed to Accept CENIC CalREN Handoff to Campus
Current UCSD Experimental Optical Core:Ready to Couple to CENIC L1, L2, L3 Services
Source: Phil Papadopoulos, SDSC/Calit2 (Quartzite PI, OptIPuter co-PI)
Funded by NSF MRI
Grant
Lucent
Glimmerglass
Force10
OptIPuter Border Router
CENIC L1, L2Services
Cisco 6509
Goals by 2008:
>= 50 endpoints at 10 GigE
>= 32 Packet switched
>= 32 Switched wavelengths
>= 300 Connected endpoints
Approximately 0.5 TBit/s Arrive at the “Optical” Center
of CampusSwitching will be a Hybrid
Combination of: Packet, Lambda, Circuit --OOO and Packet Switches
Already in Place
Planned UCSD Production Campus Cyberinfrastructure Supporting Data Intensive Biomedical Research
Source: Phil Papadopoulos, SDSC/Calit2; Elazar Harel, UCSD
N x 10 GbitN x 10 Gbit
10 Gigabit L2/L3 Switch
Eco-Friendly Storage and
Compute
Microarray
Your Lab Here
Active Data Replication
Wide-Area 10G• CENIC/HPRng• NLR Cavewave• I2 NewNet• Cinegrid• …
On-Demand Physical
Connections
“Network in a box”• > 200 Connections• DWDM or Gray Optics
N x 10 Gbit
Single 10 Gbit
Mass SpecMicroscopes
Data Repositories
UCSD Planned Optical NetworkedBiomedical Researchers and Instruments
Cellular & Molecular Medicine West
National Center for
Microscopy & Imaging
Biomedical Research
Center for Molecular Genetics Pharmaceutical
Sciences Building
Cellular & Molecular Medicine East
CryoElectron Microscopy Facility
Radiology Imaging Lab
Bioengineering
Calit2@UCSD
San Diego Supercomputer
Center
• Connects at 10 Gbps :– Microarrays
– Genome Sequencers
– Mass Spectrometry
– Light and Electron Microscopes
– Whole Body Imagers
– Computing
– Storage
Calit2/SDSC Proposal to Create a UC Cyberinfrastructure
of OptIPuter “On-Ramps” to TeraGrid Resources
UC San Francisco
UC San Diego
UC Riverside
UC Irvine
UC Davis
UC Berkeley
UC Santa Cruz
UC Santa Barbara
UC Los Angeles
UC Merced
OptIPuter + CalREN-XD + TeraGrid = “OptiGrid”
Source: Fran Berman, SDSC , Larry Smarr, Calit2
Creating a Critical Mass of End Users on a Secure LambdaGrid
Created 09-27-2005 by Garrett Hildebrand
Modified 02-28-2006 by Smarr/Hildebrand
Calit2 Building
UCInet
10 GE
HIPerWall
LosAngeles
SPDS
Catalyst 3750 in CSI
ONS 15540 WDM at UCI campus MPOE (CPL)
1 GE DWDM Network Line Tustin CENIC CalREN
POP
UCSD Optiputer Network
10 GE DWDM Network Line
Engineering Gateway Building,
Catalyst 3750 in 1st floor IDF
Catalyst 6500,
1st floor MDF
Wave-2: layer-2 GE. 67.58.33.0/25 using 11-126 at UCI. GTWY is .1
Floor 2 Catalyst 6500
Floor 3 Catalyst 6500
Floor 4 Catalyst 6500
Wave-1: layer-2 GE 67.58.21.128/25 UCI using 141-254. GTWY .128
ESMF
Catalyst 3750 in NACS Machine Room (Optiputer)
Kim JitterMeasurements Lab E1127
Wave 1 1GE
Wave 2 1GE
OptIPuter@UCI is Up and Working
Berns’ Lab--Remote Microscopy
Beckman Laser Institute Bldg.
Creating a Digital MooreaCalit2 Collaboration with UC Gump Station (UCB, UCSB)
Calit2 ReefBot Design for Digital Reef Mapping
Deck covered with solar photovoltaic
collector
Flotation ball to prevent capsize +
RADAR retro-reflector
2.2 KW Diesel Generator set
Video camera for forward looking
navigation
Sealed instrumentation &
control module
Mast includes: air intake for engine +
antenna
360 degree azipod propulsion with weed shedding prop and
complete guarding.
Basic hull: Inflatable pontoons on sides with rigid aluminum center
section.
4 deep-cycle marine batteries for energy
storage
WiFi Radioto Send Data to
Shore
UC Systems Collaborating in Studies Integrating Brain Imaging and Genetics
• Institutional Research Program: Imaging Genetics• UCI, UCLA, UCSD, UCSF, UCDavis• Transdisciplinary: molecular to clinical• Collaborate with CalIT2 for infrastructure
Source: Steven Potkin, UCI