From Jisc's campus network engineering for data-intensive science workshop on 19 October 2016.
https://www.jisc.ac.uk/events/campus-network-engineering-for-data-intensive-science-workshop-19-oct-2016
Solving Network Throughput Problems at the Diamond Light SourceJisc
From Jisc's campus network engineering for data-intensive science workshop on 19 October 2016.
https://www.jisc.ac.uk/events/campus-network-engineering-for-data-intensive-science-workshop-19-oct-2016
From Jisc's campus network engineering for data-intensive science workshop on 19 October 2016.
https://www.jisc.ac.uk/events/campus-network-engineering-for-data-intensive-science-workshop-19-oct-2016
This document discusses the concept of a Science DMZ, which consists of three key components: 1) a dedicated "friction-free" network path with high-performance networking devices located near the site perimeter to facilitate science data transfer, 2) dedicated high-performance data transfer nodes optimized for data transfer tools, and 3) a performance measurement/test node. It contrasts this approach with the typical ad-hoc deployment of a data transfer node wherever space allows, which often fails to provide necessary performance. Details of an example Science DMZ deployment at Lawrence Berkeley National Laboratory are provided.
From Jisc's campus network engineering for data-intensive science workshop on 19 October 2016.
https://www.jisc.ac.uk/events/campus-network-engineering-for-data-intensive-science-workshop-19-oct-2016
Goonhilly Earth Station played a key role in the development of the Internet. It was involved in the first demonstration of packet radio networking across the Atlantic in 1977. Goonhilly is now exploring ways to extend Internet connectivity to space, such as by developing disruption tolerant networking to enable an interplanetary Internet and supporting private lunar missions. Goonhilly also aims to diversify its business by offering commercial satellite services and partnering with universities on radio astronomy research.
Shared services - the future of HPC and big data facilities for UK researchMartin Hamilton
Slides from Jisc panel session at HPC & Big Data 2016 with contributions from the Francis Crick Institute, QMUL and King's College London covering their use of the Jisc shared data centre and the eMedLab project
The user -driven evolution of Janet - Jisc Digifest 2016Jisc
Janet is one of the world’s most advanced networks built to support research and education across the UK, and through participation in GÉANT provides global reach, supporting key activities such as transnational education and access to global research facilities.
The latest version of the network – Janet6 – came into operation in November 2013. This talk will take a look at user requirements and how these are shaping the continued evolution of Janet to ensure that a flexible, reliable and secure network service is provided.
Solving Network Throughput Problems at the Diamond Light SourceJisc
From Jisc's campus network engineering for data-intensive science workshop on 19 October 2016.
https://www.jisc.ac.uk/events/campus-network-engineering-for-data-intensive-science-workshop-19-oct-2016
From Jisc's campus network engineering for data-intensive science workshop on 19 October 2016.
https://www.jisc.ac.uk/events/campus-network-engineering-for-data-intensive-science-workshop-19-oct-2016
This document discusses the concept of a Science DMZ, which consists of three key components: 1) a dedicated "friction-free" network path with high-performance networking devices located near the site perimeter to facilitate science data transfer, 2) dedicated high-performance data transfer nodes optimized for data transfer tools, and 3) a performance measurement/test node. It contrasts this approach with the typical ad-hoc deployment of a data transfer node wherever space allows, which often fails to provide necessary performance. Details of an example Science DMZ deployment at Lawrence Berkeley National Laboratory are provided.
From Jisc's campus network engineering for data-intensive science workshop on 19 October 2016.
https://www.jisc.ac.uk/events/campus-network-engineering-for-data-intensive-science-workshop-19-oct-2016
Goonhilly Earth Station played a key role in the development of the Internet. It was involved in the first demonstration of packet radio networking across the Atlantic in 1977. Goonhilly is now exploring ways to extend Internet connectivity to space, such as by developing disruption tolerant networking to enable an interplanetary Internet and supporting private lunar missions. Goonhilly also aims to diversify its business by offering commercial satellite services and partnering with universities on radio astronomy research.
Shared services - the future of HPC and big data facilities for UK researchMartin Hamilton
Slides from Jisc panel session at HPC & Big Data 2016 with contributions from the Francis Crick Institute, QMUL and King's College London covering their use of the Jisc shared data centre and the eMedLab project
The user -driven evolution of Janet - Jisc Digifest 2016Jisc
Janet is one of the world’s most advanced networks built to support research and education across the UK, and through participation in GÉANT provides global reach, supporting key activities such as transnational education and access to global research facilities.
The latest version of the network – Janet6 – came into operation in November 2013. This talk will take a look at user requirements and how these are shaping the continued evolution of Janet to ensure that a flexible, reliable and secure network service is provided.
The STARS Shared Initiative - Pablo de Castro and Jackie ProvenRepository Fringe
The STARS Shared Initiative is a joint initiative between the University of St Andrews, RepositoryNet, and SDL to enhance the university's CRIS/IR research information management system. The initiative will test RepositoryNet's repository service catalogue by implementing services like SWORD, IRUS-UK, OpenAIRE, and ORCID in both the university's Pure CRIS and DSpace-based institutional repository. This will provide an example of delivering repository services in a complex higher education repository landscape and reinforce the University of St Andrews' research information priorities.
Cloud for Research and Innovation - UK USA HPC workshop, Oxford, July 205Martin Hamilton
How can public cloud and technologies like Docker and OpenStack help to deliver next generation scientific computing infrastructure? My talk for the UK/USA HPC workshop in July 2015, organized by HPC-SIG (UK) and CASC (USA).
This document provides an overview of Bionimbus and the Open Cloud Consortium (OCC). Bionimbus is an open source cloud for biomedical research that provides services like elastic computing, databases, data transport and analysis pipelines. The OCC operates open clouds and develops standards to bridge private and public clouds. It runs an Open Cloud Testbed and is working to build an Open Science Data Cloud. The OCC aims to develop interoperable cloud architectures and operate infrastructure at data center scale to support open science.
UCL Computer Science is planning experimental tests of their new 100G Ethernet network hardware over the next six months. The hardware only arrived in late June 2018. Initial tests will connect two hosts at 100G and test Mellanox InfiniBand application performance. Future tests are planned for UCL CS's external 100G connection, CS HEN network, and cluster storage. The department aims to push the boundaries of high performance networking and storage.
Grid'5000: Running a Large Instrument for Parallel and Distributed Computing ...Frederic Desprez
The increasing complexity of available infrastructures (hierarchical, parallel, distributed, etc.) with specific features (caches, hyper-threading, dual core, etc.) makes it extremely difficult to build analytical models that allow for a satisfying prediction. Hence, it raises the question on how to validate algorithms and software systems if a realistic analytic study is not possible. As for many other sciences, the one answer is experimental validation. However, such experimentations rely on the availability of an instrument able to validate every level of the software stack and offering different hardware and software facilities about compute, storage, and network resources.
Almost ten years after its premises, the Grid'5000 testbed has become one of the most complete testbed for designing or evaluating large-scale distributed systems. Initially dedicated to the study of large HPC facilities, Grid’5000 has evolved in order to address wider concerns related to Desktop Computing, the Internet of Services and more recently the Cloud Computing paradigm. We now target new processors features such as hyperthreading, turbo boost, and power management or large applications managing big data. In this keynote we will both address the issue of experiments in HPC and computer science and the design and usage of the Grid'5000 platform for various kind of applications.
In this presentation from the Dell booth at SC13, Joseph Antony from NCI describes how they are using HPC Virtualization to meet user needs.
Watch the video presentation: http://insidehpc.com/2013/12/05/panel-discussion-thought-hpc-virtualization-never-going-happen/
The document describes several deployment scenarios for storing and processing scientific data from instruments like the MAGIC Telescopes.
It outlines scenarios for: 1) large file safe-keeping, 2) mixed file safe-keeping including smaller files and reprocessing outputs, 3) in-archive data processing, 4) distributing data to instrument analysts, and 5) external user access.
Challenges include meeting data transfer timelines, ensuring data integrity over long periods, and providing flexible access and metadata tools to support analysis and discovery. Commercial providers could offer solutions but managing trust and customization needs is also discussed.
The document discusses requirements for archiving data from three particle accelerators and experimental areas at DESY: PETRA III, FLASH, and EuXFEL. It describes the large volumes of data generated, currently around 30 petabytes stored across various systems. A new archive system is proposed to seamlessly integrate with existing workflows and storage systems. The system should support a variety of users from individual scientists to large collaborations, with options for automated and manual archiving of experiments and analysis results over durations of 5-10 years.
Supercomputing and the cloud - the next big paradigm shift?Martin Hamilton
How can cloud technologies help us to address the challenges of re-use of research data and software and reproducibility of experiments? My slides from the University of Birmingham BEARcloud launch event, October 2016
In 2001, as early high-speed networks were deployed, George Gilder observed that “when the network is as fast as the computer's internal links, the machine disintegrates across the net into a set of special purpose appliances.” Two decades later, our networks are 1,000 times faster, our appliances are increasingly specialized, and our computer systems are indeed disintegrating. As hardware acceleration overcomes speed-of-light delays, time and space merge into a computing continuum. Familiar questions like “where should I compute,” “for what workloads should I design computers,” and "where should I place my computers” seem to allow for a myriad of new answers that are exhilarating but also daunting. Are there concepts that can help guide us as we design applications and computer systems in a world that is untethered from familiar landmarks like center, cloud, edge? I propose some ideas and report on experiments in coding the continuum.
High Performance Computing in the Cloud is viable in numerous use cases. Common to all successful use cases for cloud-based HPC is the ability embrace latency. Not surprisingly then, early successes were achieved with embarrassingly parallel HPC applications involving minimal amounts of data - in other words, there was little or no latency to be hidden. Over the fulness of time, however, the HPC-cloud community has become increasingly adept in its ability to ‘hide’ latency and, in the process, support increasingly more sophisticated HPC use cases in public and private clouds. Real-world use cases, deemed relevant to remote sensing, will illustrate aspects of these sophistications for hiding latency in accounting for large volumes of data, the need to pass messages between simultaneously executing components of distributed-memory parallel applications, as well as (processing) workflows/pipelines. Finally, the impact of containerizing HPC for the cloud will be considered through the relatively recent creation of the Cloud Native Computing Foundation.
Reading lists as open data - Meeting the Reading List Challenge 2016Martin Hamilton
1. The document discusses an open data project involving Jisc, Universities UK, and the Open Data Institute to make university reading lists openly available.
2. The project aims to collaborate across universities to publish reading list data in order to power applications like a book recommendation app and identify popular texts for potential deals.
3. Next steps could include using the consolidated open reading list data to recommend new texts, identify books to remove from lists, and monitor adoption of open textbooks between institutions. Barriers to sharing may include lack of common data standards.
Hybrid Cloud for CERN
Experience with Open Telecom Cloud and OpenStack
1) CERN uses a hybrid cloud approach combining on-premise resources with public clouds like the Open Telecom Cloud to address its increasing computing needs for projects like the LHC.
2) A pilot project using the Open Telecom Cloud was successful but identified some issues around networking and storage.
3) CERN is now participating in the HNSciCloud project to jointly procure innovative hybrid cloud services that fully integrate commercial clouds with in-house and European e-infrastructure resources.
This document provides an overview of Bionimbus and the Open Cloud Consortium. Bionimbus is a cloud platform for genomic data that provides services for data ingestion, storage, analysis, and sharing. It uses virtual machines and elastic cloud resources to enable large-scale genomic analyses. The Open Cloud Consortium is a non-profit focused on developing open cloud standards and operating shared cloud infrastructures for research. Its goals include bridging public and private clouds and operating large testbeds for cloud evaluation. The Consortium includes universities, companies, and government agencies working on cloud specifications, benchmarks, and shared testbed resources.
The document summarizes Dr. Larry Smarr's presentation on the Pacific Research Platform (PRP) and its role in working toward a national research platform. It describes how PRP has connected research teams and devices across multiple UC campuses for over 15 years. It also details PRP's innovations like Flash I/O Network Appliances (FIONAs) and use of Kubernetes to manage distributed resources. Finally, it outlines opportunities to further integrate PRP with the Open Science Grid and expand the platform internationally through partnerships.
A “meta‑cloud” for building clouds
Build your own cloud on our hardware resources
Agnostic to specific cloud software
Run existing cloud software stacks (like OpenStack, Hadoop, etc.)
... or new ones built from the ground up
Control and visibility all the way to the bare metal
“Sliceable” for multiple, isolated experiments at once
This document outlines principles for open transport data. It argues that publicly funded transport data should be released openly to developers to create jobs and increase transparency. Five key principles are discussed: 1) public money means public access to raw data; 2) open data should be free to use; 3) public money shouldn't fund competing services; 4) all developers need equal access and documentation; 5) data should be released in its original formats. Examples of both compliant and non-compliant organizations are provided. The overall goal is to organize government to better release and reform existing transport data according to these principles.
Stream data mining & CluStream frameworkYueshen Xu
This document summarizes a framework for clustering evolving data streams. The framework uses micro-clusters to represent subsets of data points and clusters micro-clusters into macro-clusters over different time horizons. Micro-clusters are represented by cluster feature vectors (CFVs) and are updated when new data points arrive by joining, deleting, or merging micro-clusters. Macro-clusters are formed by applying a modified k-means algorithm that clusters micro-clusters represented by their CFVs over different time periods.
From Jisc's campus network engineering for data-intensive science workshop on 19 October 2016.
https://www.jisc.ac.uk/events/campus-network-engineering-for-data-intensive-science-workshop-19-oct-2016
From Jisc's campus network engineering for data-intensive science workshop on 19 October 2016.
https://www.jisc.ac.uk/events/campus-network-engineering-for-data-intensive-science-workshop-19-oct-2016
The STARS Shared Initiative - Pablo de Castro and Jackie ProvenRepository Fringe
The STARS Shared Initiative is a joint initiative between the University of St Andrews, RepositoryNet, and SDL to enhance the university's CRIS/IR research information management system. The initiative will test RepositoryNet's repository service catalogue by implementing services like SWORD, IRUS-UK, OpenAIRE, and ORCID in both the university's Pure CRIS and DSpace-based institutional repository. This will provide an example of delivering repository services in a complex higher education repository landscape and reinforce the University of St Andrews' research information priorities.
Cloud for Research and Innovation - UK USA HPC workshop, Oxford, July 205Martin Hamilton
How can public cloud and technologies like Docker and OpenStack help to deliver next generation scientific computing infrastructure? My talk for the UK/USA HPC workshop in July 2015, organized by HPC-SIG (UK) and CASC (USA).
This document provides an overview of Bionimbus and the Open Cloud Consortium (OCC). Bionimbus is an open source cloud for biomedical research that provides services like elastic computing, databases, data transport and analysis pipelines. The OCC operates open clouds and develops standards to bridge private and public clouds. It runs an Open Cloud Testbed and is working to build an Open Science Data Cloud. The OCC aims to develop interoperable cloud architectures and operate infrastructure at data center scale to support open science.
UCL Computer Science is planning experimental tests of their new 100G Ethernet network hardware over the next six months. The hardware only arrived in late June 2018. Initial tests will connect two hosts at 100G and test Mellanox InfiniBand application performance. Future tests are planned for UCL CS's external 100G connection, CS HEN network, and cluster storage. The department aims to push the boundaries of high performance networking and storage.
Grid'5000: Running a Large Instrument for Parallel and Distributed Computing ...Frederic Desprez
The increasing complexity of available infrastructures (hierarchical, parallel, distributed, etc.) with specific features (caches, hyper-threading, dual core, etc.) makes it extremely difficult to build analytical models that allow for a satisfying prediction. Hence, it raises the question on how to validate algorithms and software systems if a realistic analytic study is not possible. As for many other sciences, the one answer is experimental validation. However, such experimentations rely on the availability of an instrument able to validate every level of the software stack and offering different hardware and software facilities about compute, storage, and network resources.
Almost ten years after its premises, the Grid'5000 testbed has become one of the most complete testbed for designing or evaluating large-scale distributed systems. Initially dedicated to the study of large HPC facilities, Grid’5000 has evolved in order to address wider concerns related to Desktop Computing, the Internet of Services and more recently the Cloud Computing paradigm. We now target new processors features such as hyperthreading, turbo boost, and power management or large applications managing big data. In this keynote we will both address the issue of experiments in HPC and computer science and the design and usage of the Grid'5000 platform for various kind of applications.
In this presentation from the Dell booth at SC13, Joseph Antony from NCI describes how they are using HPC Virtualization to meet user needs.
Watch the video presentation: http://insidehpc.com/2013/12/05/panel-discussion-thought-hpc-virtualization-never-going-happen/
The document describes several deployment scenarios for storing and processing scientific data from instruments like the MAGIC Telescopes.
It outlines scenarios for: 1) large file safe-keeping, 2) mixed file safe-keeping including smaller files and reprocessing outputs, 3) in-archive data processing, 4) distributing data to instrument analysts, and 5) external user access.
Challenges include meeting data transfer timelines, ensuring data integrity over long periods, and providing flexible access and metadata tools to support analysis and discovery. Commercial providers could offer solutions but managing trust and customization needs is also discussed.
The document discusses requirements for archiving data from three particle accelerators and experimental areas at DESY: PETRA III, FLASH, and EuXFEL. It describes the large volumes of data generated, currently around 30 petabytes stored across various systems. A new archive system is proposed to seamlessly integrate with existing workflows and storage systems. The system should support a variety of users from individual scientists to large collaborations, with options for automated and manual archiving of experiments and analysis results over durations of 5-10 years.
Supercomputing and the cloud - the next big paradigm shift?Martin Hamilton
How can cloud technologies help us to address the challenges of re-use of research data and software and reproducibility of experiments? My slides from the University of Birmingham BEARcloud launch event, October 2016
In 2001, as early high-speed networks were deployed, George Gilder observed that “when the network is as fast as the computer's internal links, the machine disintegrates across the net into a set of special purpose appliances.” Two decades later, our networks are 1,000 times faster, our appliances are increasingly specialized, and our computer systems are indeed disintegrating. As hardware acceleration overcomes speed-of-light delays, time and space merge into a computing continuum. Familiar questions like “where should I compute,” “for what workloads should I design computers,” and "where should I place my computers” seem to allow for a myriad of new answers that are exhilarating but also daunting. Are there concepts that can help guide us as we design applications and computer systems in a world that is untethered from familiar landmarks like center, cloud, edge? I propose some ideas and report on experiments in coding the continuum.
High Performance Computing in the Cloud is viable in numerous use cases. Common to all successful use cases for cloud-based HPC is the ability embrace latency. Not surprisingly then, early successes were achieved with embarrassingly parallel HPC applications involving minimal amounts of data - in other words, there was little or no latency to be hidden. Over the fulness of time, however, the HPC-cloud community has become increasingly adept in its ability to ‘hide’ latency and, in the process, support increasingly more sophisticated HPC use cases in public and private clouds. Real-world use cases, deemed relevant to remote sensing, will illustrate aspects of these sophistications for hiding latency in accounting for large volumes of data, the need to pass messages between simultaneously executing components of distributed-memory parallel applications, as well as (processing) workflows/pipelines. Finally, the impact of containerizing HPC for the cloud will be considered through the relatively recent creation of the Cloud Native Computing Foundation.
Reading lists as open data - Meeting the Reading List Challenge 2016Martin Hamilton
1. The document discusses an open data project involving Jisc, Universities UK, and the Open Data Institute to make university reading lists openly available.
2. The project aims to collaborate across universities to publish reading list data in order to power applications like a book recommendation app and identify popular texts for potential deals.
3. Next steps could include using the consolidated open reading list data to recommend new texts, identify books to remove from lists, and monitor adoption of open textbooks between institutions. Barriers to sharing may include lack of common data standards.
Hybrid Cloud for CERN
Experience with Open Telecom Cloud and OpenStack
1) CERN uses a hybrid cloud approach combining on-premise resources with public clouds like the Open Telecom Cloud to address its increasing computing needs for projects like the LHC.
2) A pilot project using the Open Telecom Cloud was successful but identified some issues around networking and storage.
3) CERN is now participating in the HNSciCloud project to jointly procure innovative hybrid cloud services that fully integrate commercial clouds with in-house and European e-infrastructure resources.
This document provides an overview of Bionimbus and the Open Cloud Consortium. Bionimbus is a cloud platform for genomic data that provides services for data ingestion, storage, analysis, and sharing. It uses virtual machines and elastic cloud resources to enable large-scale genomic analyses. The Open Cloud Consortium is a non-profit focused on developing open cloud standards and operating shared cloud infrastructures for research. Its goals include bridging public and private clouds and operating large testbeds for cloud evaluation. The Consortium includes universities, companies, and government agencies working on cloud specifications, benchmarks, and shared testbed resources.
The document summarizes Dr. Larry Smarr's presentation on the Pacific Research Platform (PRP) and its role in working toward a national research platform. It describes how PRP has connected research teams and devices across multiple UC campuses for over 15 years. It also details PRP's innovations like Flash I/O Network Appliances (FIONAs) and use of Kubernetes to manage distributed resources. Finally, it outlines opportunities to further integrate PRP with the Open Science Grid and expand the platform internationally through partnerships.
A “meta‑cloud” for building clouds
Build your own cloud on our hardware resources
Agnostic to specific cloud software
Run existing cloud software stacks (like OpenStack, Hadoop, etc.)
... or new ones built from the ground up
Control and visibility all the way to the bare metal
“Sliceable” for multiple, isolated experiments at once
This document outlines principles for open transport data. It argues that publicly funded transport data should be released openly to developers to create jobs and increase transparency. Five key principles are discussed: 1) public money means public access to raw data; 2) open data should be free to use; 3) public money shouldn't fund competing services; 4) all developers need equal access and documentation; 5) data should be released in its original formats. Examples of both compliant and non-compliant organizations are provided. The overall goal is to organize government to better release and reform existing transport data according to these principles.
Stream data mining & CluStream frameworkYueshen Xu
This document summarizes a framework for clustering evolving data streams. The framework uses micro-clusters to represent subsets of data points and clusters micro-clusters into macro-clusters over different time horizons. Micro-clusters are represented by cluster feature vectors (CFVs) and are updated when new data points arrive by joining, deleting, or merging micro-clusters. Macro-clusters are formed by applying a modified k-means algorithm that clusters micro-clusters represented by their CFVs over different time periods.
From Jisc's campus network engineering for data-intensive science workshop on 19 October 2016.
https://www.jisc.ac.uk/events/campus-network-engineering-for-data-intensive-science-workshop-19-oct-2016
From Jisc's campus network engineering for data-intensive science workshop on 19 October 2016.
https://www.jisc.ac.uk/events/campus-network-engineering-for-data-intensive-science-workshop-19-oct-2016
Enabling efficient movement of data into & out of a high-performance analysis...Jisc
From Jisc's campus network engineering for data-intensive science workshop on 19 October 2016.
https://www.jisc.ac.uk/events/campus-network-engineering-for-data-intensive-science-workshop-19-oct-2016
From Jisc's campus network engineering for data-intensive science workshop on 19 October 2016.
https://www.jisc.ac.uk/events/campus-network-engineering-for-data-intensive-science-workshop-19-oct-2016
From Jisc's campus network engineering for data-intensive science workshop on 19 October 2016.
https://www.jisc.ac.uk/events/campus-network-engineering-for-data-intensive-science-workshop-19-oct-2016
Archiving data from Durham to RAL using the File Transfer Service (FTS)Jisc
From Jisc's campus network engineering for data-intensive science workshop on 19 October 2016.
https://www.jisc.ac.uk/events/campus-network-engineering-for-data-intensive-science-workshop-19-oct-2016
Electron Microscopy Between OPIC, Oxford and eBICJisc
From Jisc's campus network engineering for data-intensive science workshop on 19 October 2016.
https://www.jisc.ac.uk/events/campus-network-engineering-for-data-intensive-science-workshop-19-oct-2016
This document provides an update on perfSONAR network measurement tools, the IRIS and DyGIR projects, the Archipelago measurement platform, network services on TransPAC3 and ACE, and the Data Logistics Toolkit. Key points include:
- perfSONAR and OSCARS software will be used to provide monitoring and dynamic circuit services on TransPAC3 and ACE.
- The IRIS and DyGIR projects will develop monitoring and dynamic circuit software packages for international research networks.
- The Archipelago platform conducts large-scale IPv4 topology measurements from over 50 probes worldwide.
- TransPAC3 and ACE will provide high-performance connectivity between regions and dedicated infrastructure for data movement using the
A Scenario-Based Review Of IPv6 Transition ToolsTye Rausch
1) The document discusses transition tools and techniques available for ISPs, enterprise networks, and unmanaged networks to transition to IPv6 as momentum for IPv6 adoption increases.
2) It outlines various IPv6 transition scenarios these groups may face and summarizes key transition mechanisms like dual-stack, tunneling tools like 6to4 and tunnel brokers, and encapsulation techniques for ATM and MPLS networks.
3) The 6NET project aims to encourage IPv6 adoption by providing IPv6 connectivity in academic environments and applications development for mobile and secure services.
Steve Marks.
PASIG — Preservation and Archiving Special Interest Group 2015 Meeting.
https://libraries.ucsd.edu/chronopolis/pasig/agenda_2/index_agenda.html
This document summarizes an update on IPv6 activity in CERNET2 that was presented on March 5, 2015. It discusses that CERNET2 has had a pure IPv6 backbone since 2003 connecting over 600 universities. IPv6 related research and experiments are conducted on CERNET2. Traffic statistics from January 2015 show backbone traffic exceeding 40Gbps and 10Gbps in some locations. The document also discusses challenges with scaling the DNS root server system and efforts to address this through techniques like anycasting and expanding the number of root server operators.
- James Blessing is the Deputy Director of Network Architecture at Future Services. He discussed Ciena's MCP network management software, the need for automation of network provisioning through APIs, and the JiscMail NETWORK-AUTOMATION mailing list as a resource.
- The document then covered topics like Netpath services, layer 2 and 3 VPNs, network function virtualization, IPv6 adoption, the Janet end-to-end performance initiative, science DMZ principles, network performance monitoring with perfSONAR, and working with the GÉANT project.
The SKA Project - The World's Largest Streaming Data Processorinside-BigData.com
In this presentation from the 2014 HPC Advisory Council Europe Conference, Paul Calleja from University of Cambridge presents: The SKA Project - The World's Largest Streaming Data Processor.
"The Square Kilometre Array Design Studies is an international effort to investigate and develop technologies which will enable us to build an enormous radio astronomy telescope with a million square meters of collecting area."
Watch the video presentation: http://wp.me/p3RLHQ-cot
Opening Keynote Lecture
15th Annual ON*VECTOR International Photonics Workshop
Calit2’s Qualcomm Institute
University of California, San Diego
February 29, 2016
Grid optical network service architecture for data intensive applicationsTal Lavian Ph.D.
Integrated SW System Provide the “Glue”
Dynamic optical network as a fundamental Grid service in data-intensive Grid application, to be scheduled, to be managed and coordinated to support collaborative operations
From Super-computer to Super-network
In the past, computer processors were the fastest part
peripheral bottlenecks
In the future optical networks will be the fastest part
Computer, processor, storage, visualization, and instrumentation - slower "peripherals”
eScience Cyber-infrastructure focuses on computation, storage, data, analysis, Work Flow.
The network is vital for better eScience
Pacific Wave and PRP Update Big News for Big DataLarry Smarr
The Pacific Research Platform (PRP) aims to create a "Big Data freeway system" across research institutions in the western United States and Pacific region by leveraging high-bandwidth optical fiber networks. The PRP connects multiple universities and national laboratories, providing bandwidth up to 100Gbps for data-intensive science applications. Initial testing of the PRP demonstrated disk-to-disk transfer speeds exceeding 5Gbps between many sites. The PRP will be expanded with SDN/SDX capabilities to enable even higher performance for large-scale datasets from fields like astronomy, genomics, and particle physics.
"This presentation outlines CERN's approach for decoding data from IoT devices connected to
a LoRa network using Apache Kafka, Kafka Connect, and Kafka Streams, along with Kaitai Struct,
a declarative binary format parsing language used for payload decoding.
We introduce CERN's IoT LoRa network deployment, our LoRa to Kafka bridge, and the processing platform based on Kaitai Struct specifications.
This platform is designed to automatically generate payload decoders and, if necessary, database schemas.
These are then used within Kafka Streams applications to process, decode, and inject the database schema into the payload data as required.
Our focus is on offering a simple, quick, and accessible method for end-users to easily develop and integrate new IoT streaming applications able to handle any binary data format from heterogeneous sensors (multi type, multi vendor) such as proximity, temperature, structural integrity, etc."
SDN in GÉANT provides three key points:
1) GÉANT has implemented two generations of SDN testbeds to enable researchers to test novel network ideas.
2) The Network Service Interface (NSI) protocol solves the multi-domain SDN service challenge by enabling flexible, multi-domain service chaining across network domains.
3) Work is ongoing to integrate NSI into the SDN environment by adding OpenFlow constructs to the NSI topology descriptions to bridge intra-domain OpenFlow operations with inter-domain NSI provisioning.
The document discusses how the TeraGrid initiative provides US researchers with access to large scale high performance computing resources for physics research. It describes the diverse computing resources available through TeraGrid including supercomputers, clusters, and visualization resources. It provides examples of how physics domains like lattice QCD, astrophysics, and nanoscale electronic structure are using TeraGrid resources to enable large simulations and address research challenges. Training and support resources for users are also summarized.
The document discusses the goals and activities of the NLANR/MNA Measurement and Network Analysis group. The group aims to characterize high performance computing networks through two main projects: the Active Measurement Project (AMP) and the Passive Measurement and Analysis Project (PMA). AMP conducts round-trip time, packet loss, topology and throughput measurements between over 125 monitors. PMA passively captures network traffic to study workload profiles up to OC48 speeds. The group also collaborates internationally by placing monitors in other countries and working with research networks around the world.
This document summarizes a presentation on supporting data intensive applications. It discusses the Janet end-to-end performance initiative which aims to engage with data intensive research communities to help optimize performance. Some key points include:
- Seeing increasing data intensive science applications and remote computation scenarios requiring high bandwidth.
- Importance of understanding researcher requirements and setting expectations on practical throughput limits.
- Using perfSONAR to measure network characteristics and identify performance issues between sites on the Janet network.
- Adopting the "Science DMZ" model of separating research and campus traffic to avoid bottlenecks and optimize data transfer performance.
The document discusses OCRE validation and testing. It describes the need for a tool to automate deployment of workloads for benchmarking and validating cloud services. The approach taken was to create an automated, abstracted, heterogeneous test suite using open technologies like Terraform, Ansible, Docker and Kubernetes. A variety of workloads were deployed across different domains like compute, networking, storage and machine learning. Results were collected and shared confidentially with cloud providers to evaluate areas like infrastructure, connectivity and APIs. Next steps include expanding the test catalogue and custom workload support.
Presentation of the Conference paper: "Empirical Analysis of IPv6 Transition Technologies Using the IPv6 Network Evaluation Testbed" in Tridentcom 2014, Guangzhou , China
Presentation given by PhD student Paul Harris from the Communication Systems & Networks Research Group at the University of Bristol at IEEE Globecom in Washington D.C., USA on 5th December 2016.
Scaling People, Not Just Systems, to Take On Big Data ChallengesMatthew Vaughn
Here, I describe how the Texas Advanced Computing Center has shifted its focus from traditional modeling and simulation towards fully embracing big data analytics performed by users with diverse technical backgrounds.
The Pacific Research Platform: a Science-Driven Big-Data Freeway SystemLarry Smarr
The Pacific Research Platform (PRP) is a multi-institutional partnership that establishes a high-capacity "big data freeway system" spanning the University of California campuses and other research universities in California to facilitate rapid data access and sharing between researchers and institutions. Fifteen multi-campus application teams in fields like particle physics, astronomy, earth sciences, biomedicine, and visualization drive the technical design of the PRP over five years. The goal of the PRP is to extend campus "Science DMZ" networks to allow high-speed data movement between research labs, supercomputer centers, and data repositories across campus, regional
Similar to perfSONAR: getting telemetry on your network (20)
The document announces a community launch event for digital storytelling in January 2024. It discusses using digital storytelling in higher education to support learning and teaching. Examples include using digital stories for formative assessment, reflective exercises, and research dissemination across various disciplines. Feedback from students and staff who participated in digital storytelling workshops was very positive and found it to be transformative and help give voice to their experiences. The document also profiles speakers who will discuss using digital stories to explore difficult concepts, hear the student voice, and facilitate staff reflections. It emphasizes that digital storytelling can introduce humanity and creativity into pedagogy and help develop core skills. Attendees will participate in a Miro activity to discuss benefits, applications,
This document summarizes a Jisc strategy forum that took place in Northern Ireland on December 14, 2023. It outlines Jisc's planned services and initiatives for 2023-2024, including expanding network access and launching new cybersecurity, analytics, and equipment services. It discusses feedback received from further and higher education members on how Jisc can better deliver solutions, empower communities, and provide vision/strategy. Activities at the forum focused on understanding members' needs/challenges and discussing how Jisc can better support key priorities in Northern Ireland, such as affordable infrastructure, digital skills, and cybersecurity for FE and efficiency, student experience, and collaboration for HE.
This document summarizes a Jisc Scotland strategy forum that took place on December 12, 2023. It outlines Jisc's planned solutions and services for 2023-2024 including deploying resilient Janet access, IT health checks, online surveys, SD-WAN services, and more. The document discusses how Jisc engages stakeholders through relationship management, research, communities, training and events. It summarizes feedback from further education and higher education members on how Jisc can improve advocacy by delivering the right solutions, empowering communities, and having a clear vision and strategy. Finally, it outlines activities for the forum, including understanding members' needs and priorities and discussing how Jisc supports national priorities in Scotland.
The Jisc provided a strategic update to stakeholders. Key highlights included:
- Achievements from the last year like data collection and analysis following the HESA merger, digital transformation support, and cost savings from licensing deals.
- Customer testimonials from Bridgend College on extending eduroam and from the University of Northampton on curriculum design support from Jisc.
- Priorities for the coming year like connectivity upgrades, new cybersecurity services, and improved customer experience.
- A financial summary showing income sources like membership fees and expenditures on areas like connectivity and cybersecurity.
This document summarizes VirtualSpeech, a company that provides virtual reality (VR) and artificial intelligence (AI) powered professional development training. It offers over 150 online courses covering topics like public speaking, leadership, and sales. Users can practice skills in immersive VR scenarios and receive feedback from conversational AI. The training is used by over 450,000 individuals across 130 countries and 150 universities. VirtualSpeech aims to enhance traditional learning with interactive VR practice sessions and real-time feedback to boost skills retention.
TrustArc Webinar - 2024 Global Privacy SurveyTrustArc
How does your privacy program stack up against your peers? What challenges are privacy teams tackling and prioritizing in 2024?
In the fifth annual Global Privacy Benchmarks Survey, we asked over 1,800 global privacy professionals and business executives to share their perspectives on the current state of privacy inside and outside of their organizations. This year’s report focused on emerging areas of importance for privacy and compliance professionals, including considerations and implications of Artificial Intelligence (AI) technologies, building brand trust, and different approaches for achieving higher privacy competence scores.
See how organizational priorities and strategic approaches to data security and privacy are evolving around the globe.
This webinar will review:
- The top 10 privacy insights from the fifth annual Global Privacy Benchmarks Survey
- The top challenges for privacy leaders, practitioners, and organizations in 2024
- Key themes to consider in developing and maintaining your privacy program
UiPath Test Automation using UiPath Test Suite series, part 5DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 5. In this session, we will cover CI/CD with devops.
Topics covered:
CI/CD with in UiPath
End-to-end overview of CI/CD pipeline with Azure devops
Speaker:
Lyndsey Byblow, Test Suite Sales Engineer @ UiPath, Inc.
Climate Impact of Software Testing at Nordic Testing DaysKari Kakkonen
My slides at Nordic Testing Days 6.6.2024
Climate impact / sustainability of software testing discussed on the talk. ICT and testing must carry their part of global responsibility to help with the climat warming. We can minimize the carbon footprint but we can also have a carbon handprint, a positive impact on the climate. Quality characteristics can be added with sustainability, and then measured continuously. Test environments can be used less, and in smaller scale and on demand. Test techniques can be used in optimizing or minimizing number of tests. Test automation can be used to speed up testing.
UiPath Test Automation using UiPath Test Suite series, part 6DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 6. In this session, we will cover Test Automation with generative AI and Open AI.
UiPath Test Automation with generative AI and Open AI webinar offers an in-depth exploration of leveraging cutting-edge technologies for test automation within the UiPath platform. Attendees will delve into the integration of generative AI, a test automation solution, with Open AI advanced natural language processing capabilities.
Throughout the session, participants will discover how this synergy empowers testers to automate repetitive tasks, enhance testing accuracy, and expedite the software testing life cycle. Topics covered include the seamless integration process, practical use cases, and the benefits of harnessing AI-driven automation for UiPath testing initiatives. By attending this webinar, testers, and automation professionals can gain valuable insights into harnessing the power of AI to optimize their test automation workflows within the UiPath ecosystem, ultimately driving efficiency and quality in software development processes.
What will you get from this session?
1. Insights into integrating generative AI.
2. Understanding how this integration enhances test automation within the UiPath platform
3. Practical demonstrations
4. Exploration of real-world use cases illustrating the benefits of AI-driven test automation for UiPath
Topics covered:
What is generative AI
Test Automation with generative AI and Open AI.
UiPath integration with generative AI
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Building Production Ready Search Pipelines with Spark and MilvusZilliz
Spark is the widely used ETL tool for processing, indexing and ingesting data to serving stack for search. Milvus is the production-ready open-source vector database. In this talk we will show how to use Spark to process unstructured data to extract vector representations, and push the vectors to Milvus vector database for search serving.
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
HCL Notes and Domino License Cost Reduction in the World of DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-and-domino-license-cost-reduction-in-the-world-of-dlau/
The introduction of DLAU and the CCB & CCX licensing model caused quite a stir in the HCL community. As a Notes and Domino customer, you may have faced challenges with unexpected user counts and license costs. You probably have questions on how this new licensing approach works and how to benefit from it. Most importantly, you likely have budget constraints and want to save money where possible. Don’t worry, we can help with all of this!
We’ll show you how to fix common misconfigurations that cause higher-than-expected user counts, and how to identify accounts which you can deactivate to save money. There are also frequent patterns that can cause unnecessary cost, like using a person document instead of a mail-in for shared mailboxes. We’ll provide examples and solutions for those as well. And naturally we’ll explain the new licensing model.
Join HCL Ambassador Marc Thomas in this webinar with a special guest appearance from Franz Walder. It will give you the tools and know-how to stay on top of what is going on with Domino licensing. You will be able lower your cost through an optimized configuration and keep it low going forward.
These topics will be covered
- Reducing license cost by finding and fixing misconfigurations and superfluous accounts
- How do CCB and CCX licenses really work?
- Understanding the DLAU tool and how to best utilize it
- Tips for common problem areas, like team mailboxes, functional/test users, etc
- Practical examples and best practices to implement right away
Maruthi Prithivirajan, Head of ASEAN & IN Solution Architecture, Neo4j
Get an inside look at the latest Neo4j innovations that enable relationship-driven intelligence at scale. Learn more about the newest cloud integrations and product enhancements that make Neo4j an essential choice for developers building apps with interconnected data and generative AI.
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slackshyamraj55
Discover the seamless integration of RPA (Robotic Process Automation), COMPOSER, and APM with AWS IDP enhanced with Slack notifications. Explore how these technologies converge to streamline workflows, optimize performance, and ensure secure access, all while leveraging the power of AWS IDP and real-time communication via Slack notifications.
Unlock the Future of Search with MongoDB Atlas_ Vector Search Unleashed.pdfMalak Abu Hammad
Discover how MongoDB Atlas and vector search technology can revolutionize your application's search capabilities. This comprehensive presentation covers:
* What is Vector Search?
* Importance and benefits of vector search
* Practical use cases across various industries
* Step-by-step implementation guide
* Live demos with code snippets
* Enhancing LLM capabilities with vector search
* Best practices and optimization strategies
Perfect for developers, AI enthusiasts, and tech leaders. Learn how to leverage MongoDB Atlas to deliver highly relevant, context-aware search results, transforming your data retrieval process. Stay ahead in tech innovation and maximize the potential of your applications.
#MongoDB #VectorSearch #AI #SemanticSearch #TechInnovation #DataScience #LLM #MachineLearning #SearchTechnology
Unlocking Productivity: Leveraging the Potential of Copilot in Microsoft 365, a presentation by Christoforos Vlachos, Senior Solutions Manager – Modern Workplace, Uni Systems
Cosa hanno in comune un mattoncino Lego e la backdoor XZ?Speck&Tech
ABSTRACT: A prima vista, un mattoncino Lego e la backdoor XZ potrebbero avere in comune il fatto di essere entrambi blocchi di costruzione, o dipendenze di progetti creativi e software. La realtà è che un mattoncino Lego e il caso della backdoor XZ hanno molto di più di tutto ciò in comune.
Partecipate alla presentazione per immergervi in una storia di interoperabilità, standard e formati aperti, per poi discutere del ruolo importante che i contributori hanno in una comunità open source sostenibile.
BIO: Sostenitrice del software libero e dei formati standard e aperti. È stata un membro attivo dei progetti Fedora e openSUSE e ha co-fondato l'Associazione LibreItalia dove è stata coinvolta in diversi eventi, migrazioni e formazione relativi a LibreOffice. In precedenza ha lavorato a migrazioni e corsi di formazione su LibreOffice per diverse amministrazioni pubbliche e privati. Da gennaio 2020 lavora in SUSE come Software Release Engineer per Uyuni e SUSE Manager e quando non segue la sua passione per i computer e per Geeko coltiva la sua curiosità per l'astronomia (da cui deriva il suo nickname deneb_alpha).
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-und-domino-lizenzkostenreduzierung-in-der-welt-von-dlau/
DLAU und die Lizenzen nach dem CCB- und CCX-Modell sind für viele in der HCL-Community seit letztem Jahr ein heißes Thema. Als Notes- oder Domino-Kunde haben Sie vielleicht mit unerwartet hohen Benutzerzahlen und Lizenzgebühren zu kämpfen. Sie fragen sich vielleicht, wie diese neue Art der Lizenzierung funktioniert und welchen Nutzen sie Ihnen bringt. Vor allem wollen Sie sicherlich Ihr Budget einhalten und Kosten sparen, wo immer möglich. Das verstehen wir und wir möchten Ihnen dabei helfen!
Wir erklären Ihnen, wie Sie häufige Konfigurationsprobleme lösen können, die dazu führen können, dass mehr Benutzer gezählt werden als nötig, und wie Sie überflüssige oder ungenutzte Konten identifizieren und entfernen können, um Geld zu sparen. Es gibt auch einige Ansätze, die zu unnötigen Ausgaben führen können, z. B. wenn ein Personendokument anstelle eines Mail-Ins für geteilte Mailboxen verwendet wird. Wir zeigen Ihnen solche Fälle und deren Lösungen. Und natürlich erklären wir Ihnen das neue Lizenzmodell.
Nehmen Sie an diesem Webinar teil, bei dem HCL-Ambassador Marc Thomas und Gastredner Franz Walder Ihnen diese neue Welt näherbringen. Es vermittelt Ihnen die Tools und das Know-how, um den Überblick zu bewahren. Sie werden in der Lage sein, Ihre Kosten durch eine optimierte Domino-Konfiguration zu reduzieren und auch in Zukunft gering zu halten.
Diese Themen werden behandelt
- Reduzierung der Lizenzkosten durch Auffinden und Beheben von Fehlkonfigurationen und überflüssigen Konten
- Wie funktionieren CCB- und CCX-Lizenzen wirklich?
- Verstehen des DLAU-Tools und wie man es am besten nutzt
- Tipps für häufige Problembereiche, wie z. B. Team-Postfächer, Funktions-/Testbenutzer usw.
- Praxisbeispiele und Best Practices zum sofortigen Umsetzen
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAU
perfSONAR: getting telemetry on your network
1. Duncan Rand, Jisc and Imperial College London
perfSONAR: getting telemetry on your network
2. WLCG/GridPP as an example community
19 Oct 2016
» TheWorldwide Large Hadron Collider Computing Grid
(WLCG) is a global collaboration of more than 170 computing
centres in 42 countries
» Its mission is to provide global computing resources to store,
distribute and analyse the ~30 petabytes of data generated
per year by the LHC experiments
» GridPP is a collaboration providing data-intensive distributed
computing resources for the UK HEP community and the UK
contribution to theWLCG
» Hierarchically arranged with four tiers:
› Tier-0 at CERN (andWigner in Hungary)
› 13Tier-1s (mainly national physics laboratories)
› 149Tier-2s (generally university physics laboratories)
› Tier-3s
3. 19 Oct 2016
» Initial modelling of LHC computing requirements suggested a
hierarchical tier-based data management and transfer model
» Data exported fromTier-0 at CERN to eachTier-1 and then on
toTier-2s
» However better than expected network bandwidth means
that the LHC experiments have been able to relax this
hierarchy
» Now data is transferred in an all-to-all mesh configuration
» Data often transferred across multiple domains
› e.g. a CMS transfer to Imperial College London might come
predominately from Fermilab near Chicago along with
other CMS sites
» So good network is crucial to the operation of theWLCG and
that means good monitoring
5. perfSONAR
19 Oct 2016
»Network monitoring tool developed by ESnet, GEANT,
Indiana University and Internet2
»'perfSONAR is a widely-deployed test and measurement
infrastructure that is used by science networks and
facilities around the world to monitor and ensure
network performance.’
»'perfSONAR’s purpose is to aid in network diagnosis by
allowing users to characterize and isolate problems. It
provides measurements of network performance metrics
over time as well as “on-demand” tests’
»http://www.perfsonar.net/about/what-is-perfsonar/
17. MaDDash visualisation dashboard
19 Oct 2016
»With large meshes it is difficult to check all hosts
»Centralised dashboards really help visualise overall
performance
»MaDDash (Monitoring and Debugging Dashboard)
displays meshes of perfSONAR hosts
»Many examples of MaDDash dashboards, e.g. ICNRG,
WLCG
»WLCG dashboard has two aspects
› Open Monitoring Distribution (Nagios monitoring)
› MaDDash
»http://psmad.grid.iu.edu/maddash-webui/
18. perfSONAR configuration interface
19 Oct 2016
»A perfSONAR host can participate in multiple meshes
»Configuration interface and auto-URL enables dynamic
configuration of entire network
McKee et al.
CHEP2015
19. 19 Oct 2016
»Adding and removing hosts from the mesh configuration
is very simple
»Makes use of aWLCG database of hosts
»Version of GUI developed by OSG to be included in
perfSONAR toolkit
24. Dual-stack perfSONAR measurements
19 Oct 2016
»IPv6 rollout is slow but steady
»Assumption (hope) that future campus upgrades will
include provision of IPv6
»perfSONAR supports IPv4 and IPv6 measurements
»Can leave perfSONAR hosts to default to using IPv6 if it
exists but then not always clear which is in use
»Otherwise can force with "ipv6_only": "1” parameter
25. WLCG/HEPiX IPv6 Working Groups
19 Oct 2016
»TheWLCG has an ongoing effort to promote the
adoption of IPv6
»Aim to be able to allow sites to offer IPv6-only
computing resources to theWLCG by April 2017
»HEPiX/WLCG IPv6 working groups looking into issues
»Developed mesh to track roll-out of IPv6 capable
perfSONAR hosts within WLCG
»Currently twenty oneWLCG perfSONAR dual-stack
nodes are in the mesh
30. Small perfSONAR node projects
19 Oct 2016
»DataTransfer Zones need well-specified, dedicated
hardware to run perfSONAR hosts
»Requires some investment of time and money
»Would be nice to have an easier way to get any idea of
network performance
»GÉANT have developed a small perfSONAR node using
Gigabyte Brix devices costing about £150-200 each
»Using these in a short but time-limited small perfSONAR
node project
»IPv6 included from the start
GÉANT
33. Small perfSONAR node projects
19 Oct 2016
»Jisc would like to take this project forward
»Will probably use existing image
»Send out small perfSONAR node to users who wish to
get a rapid and easy idea of their network performance
»For example a scientist in a UK institute with slow
download of data set from e.g. Diamond or Jasmin
»Also plan to produce a UK mesh into which these small
nodes could be added more or less temporarily
»Training course on how to set up such a mesh being run
by GEANT in Zurich on 4th November 2016
› https://eventr.geant.org/events/2496
34. Improving diagnostics: Pundit
19 Oct 2016
»A large mesh such as those in use byWLCG contains a lot
of useful data
»Should be possible to use network tomography to, for
example, identify problematic routers by correlating
traceroute and performance data
»PUNDIT project in US aimed at this
»Additional executable installed on perfSONAR host
»More details: http://pundit.gatech.edu and
https://indico.cern.ch/event/505613/contributions/222742
8/
35. Summary
19 Oct 2016
»perfSONAR is a valuable resource for characterising and
diagnosing network performance
»Bandwidth nodes typically record throughput and
traceroute data; latency nodes record latency and loss
»Network administrators should consider installing several
at pertinent places, e.g. at the border, next to storage etc
»Meshes together with MadDash dashboards allow
relatively easy monitoring of groups of hosts
»Future perfSONAR meshes should include IPv6
»Development work is ongoing to improve the automatic
notification and diagnosis of network faults
Go to ‘View’ menu > ‘Header and Footer…’ to edit the footers on this slide (click ‘Apply’ to change only the currently selected slide, or ‘Apply to All’ to change the footers on all slides).
To add a background image to this slide; drag a picture to the placeholder or click the icon in the centre of the placeholder to browse for and add another image. Once added, the image can be cropped, resized or repositioned to suit.
IP addresses IPv4 and v6
Problems with the perfsonar host itself
Sometimes the issue is with campus network. This took a long time to resolve