Invited Remote Lecture to SC21
The International Conference for High Performance Computing, Networking, Storage, and Analysis
St. Louis, Missouri
November 18, 2021
Report
Share
Report
Share
1 of 45
More Related Content
Similar to The Rise of Supernetwork Data Intensive Computing
Supercomputers and Supernetworks are Transforming ResearchLarry Smarr
09.03.25
Invited Talk
Computing Research that Changed the World: Reflections and Perspectives
Title: Supercomputers and Supernetworks are Transforming Research
Washington, DC
The Pacific Research Platform (PRP) is a multi-institutional cyberinfrastructure project that connects researchers across California and beyond to share large datasets. It spans the 10 University of California campuses, major private research universities, supercomputer centers, and some out-of-state universities. Fifteen multi-campus research teams in fields like physics, astronomy, earth sciences, biomedicine, and multimedia will drive the technical needs of the PRP over five years. The goal is to create a "big data freeway" to allow high-speed sharing of data between research labs, supercomputers, and repositories across multiple networks without performance loss over long distances.
The document provides an introduction to the internet and the world wide web for a course on website design and development. It discusses the history of the internet and how it originated as a military network before becoming publicly available. It defines key terms like web browser, website, and homepage. It explains that the internet is the underlying infrastructure of hardware, while the world wide web is the software comprising hyperlinked web pages. The internet connects millions of computers and networks globally through protocols like TCP/IP, while the web is governed by HTTP and links files and documents stored on various computers.
- The Pacific Research Platform (PRP) interconnects campus DMZs across multiple institutions to provide high-speed connectivity for data-intensive research.
- The PRP utilizes specialized data transfer nodes called FIONAs that provide disk-to-disk transfer speeds of 10-100Gbps.
- Early applications of the PRP include distributing telescope data between UC campuses, connecting particle physics experiments to computing resources, and enabling real-time wildfire sensor data analysis.
COMPED9 Module 2 The Internet and the WebJeremy Flores
The document summarizes the history and development of the Internet and World Wide Web. It describes how JCR Licklider first proposed a global network of computers in the 1960s which led to the development of ARPANET and packet switching. The ARPANET connected four universities in 1969 which is seen as the beginning of the Internet. Key developments included the introduction of email, FTP, and TCP/IP. Tim Berners-Lee later introduced the World Wide Web in 1989 as a way to share information over the Internet using hyperlinks. The graphical web browser Mosaic popularized the Web in the early 1990s. The document distinguishes that the Internet is the global network itself while the Web is an interface that runs on
Montana State, Research Networking and the Outcomes from the First National R...Jerry Sheehan
Presentation at Educause 17 with our Partner Cisco on Research networking, covers our campus experience and the first National Research Platform Workshop findings
Activity 12 timeline history of internetCoralyn Lim
The document traces the origins and development of the Internet from 1957 to 2012. It highlights several key events and innovations such as the launch of Sputnik which led to the creation of ARPA (1957), the proposal of a "Galactic Network" by J.C.R. Licklider (1962), the development of packet switching by Paul Baran (1962), the creation of the first email program by Ray Tomlinson (1972), and the development of TCP/IP by Vint Cerf and Bob Kahn (1973) which allowed computers on different networks to communicate. The document also notes the development of technologies like Ethernet (1976), satellite networks (1976), and the world wide web (1991) which expanded
The Academic and R&D Sectors' Current and Future Broadband and Fiber Access N...Larry Smarr
05.02.23
Invited Access Grid Talk
MSCMC FORUM Series
Examining the National Vision for Global Peace and Prosperity
Title: The Academic and R&D Sectors' Current and Future Broadband and Fiber Access Needs for US Global Competitiveness
Arlington, VA
Metacomputer Architecture of the Global LambdaGridLarry Smarr
06.01.13
Invited Talk
Department of Computer Science
Donald Bren School of Information and Computer Sciences
Title: Metacomputer Architecture of the Global LambdaGrid
Irvine, CA
Vinton Cerf describes the early history and development of the Internet. The ARPANET was created in the late 1960s to enable communication between computers. In the 1970s, Cerf and others worked to develop the TCP/IP protocols to allow different networks to interconnect, laying the foundation for today's Internet. A key demonstration in 1977 showed data traveling over multiple networks between the US and Europe. Standards were finalized in the late 1970s, and the TCP/IP protocols were adopted for military use in 1980.
The document defines the internet as a global system of interconnected computer networks that transmit data via packets. It then outlines the evolution of the internet from its origins as the ARPANET in the 1960s developed by US military researchers, to the creation of TCP/IP in the 1970s which allowed different networks to communicate, and the introduction of the world wide web in the 1990s. Finally, it lists some common internet applications including email, file transfer, remote login, news, hypertext, chat, and online shopping.
The document provides a detailed timeline and history of the development of the Internet from the late 1950s to present day. It traces the early research on packet switching networks funded by ARPA/DARPA and the development of TCP/IP in the 1970s. Key events include the creation of the ARPANET in the 1960s, the World Wide Web in the early 1990s, and the commercialization of the Internet in the mid-1990s. The document also describes common Internet protocols like HTTP, DNS, and how IP addresses and domain names work.
High Performance Cyberinfrastructure for Data-Intensive ResearchLarry Smarr
This document summarizes a lecture given by Dr. Larry Smarr on high performance cyberinfrastructure for data-intensive research. The summary discusses:
1) The need for dedicated high-bandwidth networks separate from the shared internet to enable big data research due to the increasing volume of digital scientific data.
2) Extensions being made to networks like CENIC in California to provide campus "Big Data Freeways" connecting instruments, computing resources, and remote facilities.
3) The use of networks like HPWREN to provide high-performance wireless access for data-intensive applications in rural areas like astronomy, wildfire detection, and more.
The document discusses the past, present, and future of the internet and related technologies. It notes that computing power and internet connectivity have increased dramatically over time, enabling billions of devices to connect. Issues around data volume, security, applications, and societal impacts are discussed. The future internet is predicted to involve trillions of connected devices, data-driven applications, integrated physical and digital worlds, and challenges around privacy, ownership and control of data, and ensuring open access.
Design and development of a web-based data visualization software for politic...Alexandros Britzolakis
Presenting a tool for identifying political popularity over Twitter. AthPPA (which stands for Athena Political Popularity Analysis) is a tool for identifying how popular a political leader is over Twitter. For the purposes of this dissertation the Twitter accounts of the most prominent Greek political leaders have been identified. Structured data such as likes, re-tweets, text-length per tweet as well as the number of subscribers per account have been visualized. Furthermore, sentiment analysis is calculated and visualized using spaCy module and a sentiment lexicon which contains a set of emotion based labeled words.
Similar to The Rise of Supernetwork Data Intensive Computing (20)
My Remembrances of Mike Norman Over The Last 45 YearsLarry Smarr
Mike Norman has been a leader in computational astrophysics for over 45 years. Some of his influential work includes:
- Cosmic jet simulations in the early 1980s which helped explain phenomena from galactic centers.
- Pioneering the use of adaptive mesh refinement in the 1990s to achieve dynamic load balancing on supercomputers.
- Massive cosmology simulations in the late 2000s with over 100 trillion particles using thousands of processors across multiple supercomputing sites, producing petabytes of data.
- Developing end-to-end workflows in the 2000s to couple supercomputers, high-speed networks, and large visualization systems to enable real-time analysis of extremely large astrophysics simulations.
Metagenics How Do I Quantify My Body and Try to Improve its Health? June 18 2019Larry Smarr
Larry Smarr discusses quantifying his body and health over time through extensive self-tracking. He measures various biomarkers through regular blood tests and analyzes his gut microbiome by sequencing stool samples. This revealed issues like chronic inflammation and an unhealthy microbiome. Smarr then took steps like a restricted eating window and increasing plant diversity in his diet, which reversed metabolic syndrome issues and correlated with shifts in his microbiome ecology. His goal is to continue precisely measuring factors like toxins, hormones, gut permeability and food/supplement impacts to further optimize his health.
The History of Embeddings & Multimodal EmbeddingsZilliz
Frank Liu will walk through the history of embeddings and how we got to the cool embedding models used today. He'll end with a demo on how multimodal RAG is used.
UiPath Community Day Amsterdam: Code, Collaborate, ConnectUiPathCommunity
Welcome to our third live UiPath Community Day Amsterdam! Come join us for a half-day of networking and UiPath Platform deep-dives, for devs and non-devs alike, in the middle of summer ☀.
📕 Agenda:
12:30 Welcome Coffee/Light Lunch ☕
13:00 Event opening speech
Ebert Knol, Managing Partner, Tacstone Technology
Jonathan Smith, UiPath MVP, RPA Lead, Ciphix
Cristina Vidu, Senior Marketing Manager, UiPath Community EMEA
Dion Mes, Principal Sales Engineer, UiPath
13:15 ASML: RPA as Tactical Automation
Tactical robotic process automation for solving short-term challenges, while establishing standard and re-usable interfaces that fit IT's long-term goals and objectives.
Yannic Suurmeijer, System Architect, ASML
13:30 PostNL: an insight into RPA at PostNL
Showcasing the solutions our automations have provided, the challenges we’ve faced, and the best practices we’ve developed to support our logistics operations.
Leonard Renne, RPA Developer, PostNL
13:45 Break (30')
14:15 Breakout Sessions: Round 1
Modern Document Understanding in the cloud platform: AI-driven UiPath Document Understanding
Mike Bos, Senior Automation Developer, Tacstone Technology
Process Orchestration: scale up and have your Robots work in harmony
Jon Smith, UiPath MVP, RPA Lead, Ciphix
UiPath Integration Service: connect applications, leverage prebuilt connectors, and set up customer connectors
Johans Brink, CTO, MvR digital workforce
15:00 Breakout Sessions: Round 2
Automation, and GenAI: practical use cases for value generation
Thomas Janssen, UiPath MVP, Senior Automation Developer, Automation Heroes
Human in the Loop/Action Center
Dion Mes, Principal Sales Engineer @UiPath
Improving development with coded workflows
Idris Janszen, Technical Consultant, Ilionx
15:45 End remarks
16:00 Community fun games, sharing knowledge, drinks, and bites 🍻
"Building Future-Ready Apps with .NET 8 and Azure Serverless Ecosystem", Stan...Fwdays
.NET 8 brought a lot of improvements for developers and maturity to the Azure serverless container ecosystem. So, this talk will cover these changes and explain how you can apply them to your projects. Another reason for this talk is the re-invention of Serverless from a DevOps perspective as a Platform Engineering trend with Backstage and the recent Radius project from Microsoft. So now is the perfect time to look at developer productivity tooling and serverless apps from Microsoft's perspective.
Retrieval Augmented Generation Evaluation with RagasZilliz
Retrieval Augmented Generation (RAG) enhances chatbots by incorporating custom data in the prompt. Using large language models (LLMs) as judge has gained prominence in modern RAG systems. This talk will demo Ragas, an open-source automation tool for RAG evaluations. Christy will talk about and demo evaluating a RAG pipeline using Milvus and RAG metrics like context F1-score and answer correctness.
It's your unstructured data: How to get your GenAI app to production (and spe...Zilliz
So you've successfully built a GenAI app POC for your company -- now comes the hard part: bringing it to production. Aparavi addresses the challenges of AI projects while addressing data privacy and PII. Our Service for RAG helps AI developers and data scientists to scale their app to 1000s to millions of users using corporate unstructured data. Aparavi’s AI Data Loader cleans, prepares and then loads only the relevant unstructured data for each AI project/app, enabling you to operationalize the creation of GenAI apps easily and accurately while giving you the time to focus on what you really want to do - building a great AI application with useful and relevant context. All within your environment and never having to share private corporate data with anyone - not even Aparavi.
Finetuning GenAI For Hacking and DefendingPriyanka Aash
Generative AI, particularly through the lens of large language models (LLMs), represents a transformative leap in artificial intelligence. With advancements that have fundamentally altered our approach to AI, understanding and leveraging these technologies is crucial for innovators and practitioners alike. This comprehensive exploration delves into the intricacies of GenAI, from its foundational principles and historical evolution to its practical applications in security and beyond.
"Hands-on development experience using wasm Blazor", Furdak Vladyslav.pptxFwdays
I will share my personal experience of full-time development on wasm Blazor
What difficulties our team faced: life hacks with Blazor app routing, whether it is necessary to write JavaScript, which technology stack and architectural patterns we chose
What conclusions we made and what mistakes we committed
Self-Healing Test Automation Framework - HealeniumKnoldus Inc.
Revolutionize your test automation with Healenium's self-healing framework. Automate test maintenance, reduce flakes, and increase efficiency. Learn how to build a robust test automation foundation. Discover the power of self-healing tests. Transform your testing experience.
Keynote : AI & Future Of Offensive SecurityPriyanka Aash
In the presentation, the focus is on the transformative impact of artificial intelligence (AI) in cybersecurity, particularly in the context of malware generation and adversarial attacks. AI promises to revolutionize the field by enabling scalable solutions to historically challenging problems such as continuous threat simulation, autonomous attack path generation, and the creation of sophisticated attack payloads. The discussions underscore how AI-powered tools like AI-based penetration testing can outpace traditional methods, enhancing security posture by efficiently identifying and mitigating vulnerabilities across complex attack surfaces. The use of AI in red teaming further amplifies these capabilities, allowing organizations to validate security controls effectively against diverse adversarial scenarios. These advancements not only streamline testing processes but also bolster defense strategies, ensuring readiness against evolving cyber threats.
Discovery Series - Zero to Hero - Task Mining Session 1DianaGray10
This session is focused on providing you with an introduction to task mining. We will go over different types of task mining and provide you with a real-world demo on each type of task mining in detail.
Increase Quality with User Access Policies - July 2024Peter Caitens
⭐️ Increase Quality with User Access Policies ⭐️, presented by Peter Caitens and Adam Best of Salesforce. View the slides from this session to hear all about “User Access Policies” and how they can help you onboard users faster with greater quality.
Redefining Cybersecurity with AI CapabilitiesPriyanka Aash
In this comprehensive overview of Cisco's latest innovations in cybersecurity, the focus is squarely on resilience and adaptation in the face of evolving threats. The discussion covers the imperative of tackling Mal information, the increasing sophistication of insider attacks, and the expanding attack surfaces in a hybrid work environment. Emphasizing a shift towards integrated platforms over fragmented tools, Cisco introduces its Security Cloud, designed to provide end-to-end visibility and robust protection across user interactions, cloud environments, and breaches. AI emerges as a pivotal tool, from enhancing user experiences to predicting and defending against cyber threats. The blog underscores Cisco's commitment to simplifying security stacks while ensuring efficacy and economic feasibility, making a compelling case for their platform approach in safeguarding digital landscapes.
1. “The Rise of Supernetwork
Data Intensive Computing”
Invited Remote Lecture to SC21
The International Conference for High Performance
Computing, Networking, Storage, and Analysis
St. Louis, Missouri
November 18, 2021
Dr. Larry Smarr
Founding Director Emeritus, California Institute for Telecommunications and Information Technology;
Distinguished Professor Emeritus, Dept. of Computer Science and Engineering
Jacobs School of Engineering, UCSD
http://lsmarr.calit2.net
2. Abstract
Over the last 35 years, a fundamental architectural transformation in high performance data-intensive
computing has occurred, driven by the rise of optical fiber Supernetworks connecting the globe.
Ironically, this cyberinfrastructure revolution has been led by supercomputer centers, which then became
SuperNodes in this distributed system. I will review key moments, including the birth of the NSF
Supercomputer Centers and NSFnet, the gigabit testbeds, the NSF PACI program, the emergence of
Internet2 and the Regional Optical Networks, all eventually enabling, through a series of NSF grants, the
National and Global Research Platforms. Over this same period a similar cyberinfrastructure architecture
allowed the commercial clouds to develop, which are now interconnected with this academic distributed
system. Critical to this transformation has been the continual exponential rise of data and a new
generation of distributed applications utilizing this connected digital fabric. Throughout this period, the
role of the US Federal Government has been essential, anchored by the 1991 High-Performance
Computing Act, which established the Networking and Information Technology Research and
Development (NITRD) Program. Particularly important to the initiation of this distributed computing
paradigm shift was the continued visionary leadership of Representative, then Senator, then Vice
President Al Gore in the 1990s.
3. 1975-1985: My Early Research was on Computational Astrophysics
Before There Were National Academic Supercomputer Centers
I Spent a Decade Supercomputing at LLNL (with Jim Wilson) and
Then at The Max Planck Institute for Physics and Astrophysics (with Mike Norman and Karl-Heinz Winkler)
Gas Accretion Onto a Black Hole
With Wilson and Hawley
1982
Cosmic Jets Emerging From Galactic Centers
With Norman and Winkler
1981
Gravitational Radiation From Black Hole Collisions
With Eppley
1978
4. 1982-1983: Documenting The Unmet Supercomputing Needs
of a Broad Range of Disciplines Led to the NCSA Proposal to NSF
1982 1983
http://lsmarr.calit2.net/supercomputer_famine_1982.pdf http://lsmarr.calit2.net/Black_Proposal.pdf
1984: NSF Creates Office of Advanced Scientific Computing (John Connolly, Director)
Issues National Competition for Supercomputer Centers
5. 1985: NSF Adopted a DOE High-Performance Computing Model
For Two of the New NSF Supercomputer Centers
NCSA Was Modeled on LLNL SDSC Was Modeled on MFEnet
1985
6. SuperNetworks Have Co-Evolved
with Supercomputers For 35 Years
“We ought to consider
a national initiative
to build interstate highways
for information
with a fiber optics network
connecting the
major computational centers
in this country”
-Senator Al Gore
“The University of Illinois
will be experimenting with
fiber optic
"information flow pipes,"
which promise to be able
to reach
billions of bits per second.””
-NCSA Director
Larry Smarr
http://lsmarr.calit2.net/hrg-1985-tec-0068_from_1_to_806_s.pdf
1985
7. Remote Interactive Visual Supercomputing End-to-End Prototype:
Using Analog Communications to Prototype the Fiber Optic Future
“We’re using satellite technology…
to demonstrate
what It might be like to have
high-speed fiber-optic links
between advanced computers
in two different
geographic locations.”
Illinois
Boston
SIGGRAPH 1989
“What we really have to do is eliminate distance
between individuals who want to interact
with other people and computers.”
― Larry Smarr, Director, NCSA
www.youtube.com/watch?v=C3d_6lw8_0M
-Al Gore, Senator
Chair, US Senate
Subcommittee
Cray 2 Driven by
Sun Workstation
AT&T & Sun
Telepresenc
e
8. 1991: Networking Information Technology Research and Development (NITRD)
• NITRID Was Enacted in 1991 by Congress
Through the High-Performance Computing and Communication Act
• Brought Multiple Federal Agencies Together
to Plan and Coordinate Frontier Computing, Networking, Software, and Data
• Bill Was Sponsored and Driven by Senator Al Gore
December 2, 2021
9. The Bandwidth and Number of Endpoints
on NSFNET Grew Rapidly
Visualization of Inbound Traffic on the NSFNET T1 Backbone
(September 1991) by NCSA’s Donna Cox and Robert Patterson;
Data Collected by Merit Network, Inc.
1994
1991
10. • The First National 155 Mbps Research Network
– Inter-Connected Telco Networks Via IP/ATM With:
– Supercomputer Centers
– Virtual Reality Research Locations, and
– Applications Development Sites
– Into the San Diego Convention Center
– 65 Science Projects
• I-Way Featured:
– Networked Visualization Applications
– Large-Scale Immersive Displays
– I-Soft Programming Environment
– Led to the Globus Project
Supercomputing ’95:
I-WAY: A Model for Distributed Collaborative Computing
For details see:
“Overview of the I-WAY: Wide Area Visual Supercomputing”
DeFanti, Foster, Papka, Stevens, Kuhfuss
www.globus.org/sites/default/files/iway_overview.pdf SC95 Chair Sid Karin
SC95 Program Chair, Larry Smarr
11. 1990-1996 CNRI’s Gigabit Testbeds
Demonstrated Host I/O Was the Distributed Computing Bottleneck
“Host I/O proved to be
the Achilles' heel
of gigabit networking –
whereas LAN and WAN technologies were
operated in the gigabit regime, many
obstacles impeded
achieving gigabit flows
into and out of
the host computers
used in the testbeds.”
--Final Report
The Gigabit Testbed Initiative
December 1996
Corporation for
National Research Initiatives (CNRI)
Robert Kahn
CNRI Chairman, CEO & President
12. NSF’s PACI Program was Built on the vBNS
to Prototype America’s 21st Century Information Infrastructure
PACI National Technology Grid
Testbed
National Computational Science
1997
vBNS
led to
Key Role
of Miron Livny
& Condor
13. The 25 Years From the National Techology Grid
To the National Research Platform
From I-WAY to the National Technology Grid, CACM, 40, 51 (1997)
Rick Stevens, Paul Woodward, Tom DeFanti, and Charlie Catlett
14. Dave Bader Created the First Linux COTS Supercluster -Roadrunner-
on the National Technology Grid, with the Support of NCSA and NSF
NCSA Director Larry Smarr (left),
UNM President William Gordon,
and U.S. Sen. Pete Domenici
Turn on the Roadrunner Supercomputer
in April 1999
1999
National Computational Science
15. Illinois’s I-WIRE and Indiana’s I-LIGHT Dark Fiber Networks
Inspired Many Other State and Regional Optical
Source: Larry Smarr, Rick Stevens, Tom DeFanti, Charlie Catlett
1999
Today California’s
CENIC R&E
Backbone Includes
~ 8,000 Miles of
CENIC-Owned and
Managed Fiber
16. 1999: The President’s Information Technology Advisory Committee (PITAC) Report
Led to Funding NSF’s Information Technology Research (ITR) for National Priorities Program
Meeting with Vice President Gore in the White House
To Present Our PITAC Report
PITAC
Co-Chairs:
Ken Kennedy
Bill Joy
17. The OptIPuter
Exploits a New World
in Which
the Central Architectural Element
is Optical Networking,
Not Computers
to Support
Data-Intensive Scientific Research
and Collaboration
OptIPuter
NSF ITR Grant
$13.5M
PI Smarr,
Co-PIs DeFanti,
Papadopoulos, Ellisman
2002-2009
2002-2009: The NSF OptIPuter ITR Grant-
Can We Make Wide-Area Bandwidth Equal to Cluster Backplane Speeds?
18. Integrated “OptIPlatform” Cyberinfrastructure System:
A 10Gbps Lightpath Cloud
National LambdaRail
Campus
Optical
Switch
Data Repositories & Clusters
HPC
HD/4k Video Images
HD/4k Video Cams
End User
OptIPortal
10G
Lightpath
HD/4k Telepresence
Instruments
LS 2009
Slide
19. David Abramson Led OptIPuter Global Workflows and
UCSD/Monash Univ. Co-Mentoring of Undergrads and Graduate Students
First OptIPortal/Kepler
Remote Microscopy Link Feb 2009
Monash U.
UCSD
Monash U.
20. 2010-2020:
NSF Adopted a DOE High-Performance Networking Model
Science
DMZ
Data Transfer
Nodes
(DTN/FIONA)
Network
Architecture
(zero friction)
Performance
Monitoring
(perfSONAR)
ScienceDMZ Coined in 2010 by ESnet
http://fasterdata.es.net/science-dmz/
Slide Adapted From Inder Monga, ESnet
DOE
NSF
NSF Campus Cyberinfrastructure Program
2012-2020
Has Made Over 340 Awards:
Across 50 States and Territories
Slide Adapted From Kevin Thompson, NSF
21. 2013-2015: UCSD as a Laboratory for a “Big Data” 10-100 Gbps ScienceDMZ
NSF-Funded Campus CI Grants: Prism@UCSD and CHeruB
Prism@UCSD, Phil Papadopoulos, SDSC, Calit2, PI (2013-15)
CHERuB, Mike Norman, SDSC PI
CHERuB
22. (GDC)
2015 Vision: The Pacific Research Platform Will Connect Science DMZs
Creating a Regional End-to-End Science-Driven Community Cyberinfrastructure
NSF CC*DNI Grant
$6.3M 10/2015-10/2020
In Year 6 Now, Year 7 is Funded
Source: John Hess, CENIC
Supercomputer
Centers
23. PRP Website Has All Details Needed to Get Started
https://pacificresearchplatform.org/
24. 2015-2021: UCSD Designs PRP Data Transfer Nodes (DTNs) --
Flash I/O Network Appliances (FIONAs)
FIONAs Solved the the Gigabit Testbed Disk-to-Disk Data Transfer Problem
at Near Full Speed on Best-Effort 10G, 40G and 100G
FIONAs Designed by UCSD’s Phil Papadopoulos, John Graham,
Joe Keefe, and Tom DeFanti
Up to 192 TB Rotating Storage
www.pacificresearchplatform.org
Today’s
Roadrunner!
25. 2018/2019: PRP Game Changer!
Using Google’s Kubernetes to Orchestrate Containers Across the PRP
User
Applications
Containers
Clouds
26. PRP’s Nautilus Hypercluster Adopted Kubernetes
to Orchestrate Software Containers and Manage Distributed Storage
“Kubernetes with Rook/Ceph Allows Us to Manage Petabytes of
Distributed Storage and GPUs for Data Science,
While We Measure and Monitor Network Use.”
--John Graham, Calit2/QI UC San Diego
Kubernetes (K8s) is an open-source system for
automating deployment, scaling, and
management of containerized applications.
27. 2017-2020: NSF CHASE-CI Grant Adds a Machine Learning Layer
Built on Top of the Pacific Research Platform
Caltech
UCB
UCI UCR
UCSD
UCSC
Stanford
MSU
UCM
SDSU
NSF Grant for High Speed “Cloud” of 256 GPUs
For 30 ML Faculty & Their Students at 10 Campuses
for Training AI Algorithms on Big Data
PI: Larry Smarr
Co-PIs:
• Tajana Rosing
• Ken Kreutz-Delgado
• Ilkay Altintas
• Tom DeFanti
28. Original
PRP
CENIC/PW Link
2018-2021: Toward the National Research Platform (TNRP) -
Using CENIC & Internet2 to Connect Quilt Regional R&E Networks
“Towards
The NRP”
3-Year Grant
Funded
by NSF
$2.5M
October 2018
Award #1826967
PI Smarr
Co-PIs Altintas,
Papadopoulos,
Wuerthwein,
Rosing
29. Rotating Storage
4000 TB
PRP’s Nautilus is a Multi-Institution Hypercluster
Connected by Optical Networks
184 FIONAs on 25 Partner Campuses
Networked Together at 10-100Gbps
32. PRP’s Nautilus is Centered in SoCal
FIONAs
UCSD &
SDSU
UCI
Caltech
UCSB
UCR
CSUSB
33. We Measure Disk-to-Disk Throughput with 10GB File Transfer
4 Times Per Day in Both Directions for All PRP Sites
January 29,
2016
From Start of Monitoring 12 DTNs
to 24 DTNs Connected at 10-40G
in 1 ½ Years
July 21, 2017
Source: John Graham, Calit2
34. Operational Metrics: Containerized Trace Route Tool Allows Realtime Visualization
of Status of PRP Network Links on a National and Global Scale
Source: Dima Mishin, SDSC
9/16/2019
Guam
Univ. Queensland
Australia
LIGO
UK
Netherlands
Korea
36. Director:
F. Martin Ralph
Big Data Collaboration with:
Scott Sellers, PhD CHRS; Postdoc CW3E
PRP Accelerates Data-Intensive Workflow on Atmospheric Water in the West
Between NASA MERRA Archive, UC San Diego, and UC Irvine
Director:
Soroosh Sorooshian
Complete Workflow Time:
19.2 days🡪52 Minutes!
See Paper by Sellars, et al., IEEE eScience (2019)
http://lsmarr.calit2.net/sellars_accelerating_image_segmentation.pdf
37. The New Pacific Research Platform Video
Highlights 3 Different Applications Out of 600 Nautilus Namespace Projects
Pacific Research Platform Video:
www.thequilt.net/campus-cyberinfrastructure-program-resource/
www.pacificresearchplatform.org
38. Co-Existence of Interactive and
Non-Interactive Computing on PRP
GPU Simulations Needed to Improve Ice Model.
=> Results in Significant Improvement
in Pointing Resolution for Multi-Messenger Astrophysics
NSF Large-Scale Observatories Are Using PRP and OSG
as a Cohesive, Federated, National-Scale Research Data Infrastructure
NSF’s IceCube & LIGO Both See Nautilus
as Just Another OSG Resource
IceCube Used Up to Half of
PRP’s 500 GPUs in 2020!
39. UC President Napolitano's Research Catalyst Award to
UC San Diego (Tom Levy), UC Berkeley (Benjamin Porter), UC Merced (Nicola Lercari) and UCLA (Willeke Wendrich)
PRP Links At-Risk Cultural Heritage and Archaeology Datasets
to Virtual Reality Systems at Multiple Campuses
48 Megapixel CAVEkiosk
UCSD Library
48 Megapixel CAVEkiosk
UCB CITRIS Tech Museum
24 Megapixel CAVEkiosk
UCM Library
40. Once a Wildfire is Spotted, PRP Brings High-Resolution Weather Data
to Fire Modeling Workflows in WIFIRE
Real-Time
Meteorological Sensors
Weather Forecast
Landscape data
WIFIRE Firemap
Fire Perimeter
Work Flow
PRP
Source: Ilkay Altintas, SDSC
41. Community Building Though Inclusion and Diversity
• Grants
– 3 Female co-PIs
– 1 Hispanic co-PI
• Campuses
– 8 Minority-Serving Institutions in PRP/CHASE-CI
• Workshops
– NRPII Workshop Steering Committee 80% Female
– Multiple MSI, EPSCoR Focused Workshops
Jackson State University
PRP MSI Workshop
Presenting
FIONettes
42. 2021-2024 NRP Future I: Proposed Extension of Nautilus
CHASE-CI ENS, Tom DeFanti PI (NSF Award # 2120019)
CHASE-CI ABR, Larry Smarr PI (NSF Award # 2100237)
$2.8M
43. 2021-2026 NRP Future II: PRP Federates with SDSC’s EXPANSE
Using CHASE-CI Developed Composable Systems
~$20M over 5 Years
PI Mike Norman, SDSC
44. 2021-2026 NRP Future III: PRP Federates with
NSF-Funded Prototype National Research Platform
NSF Award OAC #2112167 (June 2021) [$5M Over 5 Years]
PI Frank Wuerthwein (UCSD, SDSC)
Co-PIs Tajana Rosing (UCSD), Thomas DeFanti (UCSD), Mahidhar Tatineni (SDSC), Derek Weitzel (UNL)
45. PRP/TNRP/CHASE-CI Support and Community:
• US National Science Foundation (NSF) awards to UCSD, NU, and SDSC
⮚ CNS-1456638, CNS-1730158, ACI-1540112, ACI-1541349, & OAC-1826967
⮚ OAC 1450871 (NU) and OAC-1659169 (SDSU)
• UC Office of the President, Calit2 and Calit2’s UCSD Qualcomm Institute
• San Diego Supercomputer Center and UCSD’s Research IT and Instructional IT
• Partner Campuses: UCB, UCSC, UCI, UCR, UCLA, USC, UCD, UCSB, SDSU, Caltech, NU,
UWash UChicago, UIC, UHM, CSUSB, HPWREN, UMo, MSU, NYU, UNeb, UNC,UIUC,
UTA/Texas Advanced Computing Center, FIU, KISTI, UVA, AIST
• CENIC, Pacific Wave/PNWGP, StarLight/MREN, The Quilt, Kinber, Great Plains Network,
NYSERNet, LEARN, Open Science Grid, Internet2, DOE ESnet, NCAR/UCAR & Wyoming
Supercomputing Center, AWS, Google, Microsoft, Cisco