GLIF 2018 Held at the Home of Hamlet Thumbnail
Deploy360 17 October 2018

GLIF 2018 Held at the Home of Hamlet

By Kevin MeynellFormer Senior Manager, Technical and Operational Engagement

The 18th Annual Global LambaGrid Workshop (GLIF 2018) was held on 18-21 September 2018 at the Kulturværftet in Helsingør (Elsinore), Denmark. Kronberg Castle, located next to the venue, was immortalised as Elsinore in the William Shakespeare play Hamlet, but there proved to be nothing rotten with the state of high-bandwidth networking as 50 participants from 19 countries came to hear how these networks are facilitating exascale computing in support of biological, medical, physics, energy production and environmental research, and to discuss the latest infrastructure developments.

This event was organised by myself with support from NORDUnet who hosted the event in conjunction with the 30th NORDUnet Conference (NDN18), and where I also took the opportunity to raise awareness of the MANRS initiative.

The keynote was provided by Steven Newhouse (EBI) who presented the ELIXIR Compute Platform which was being used for analysing life science data. In common with high-energy physics, genomics research produces a lot of data, but this is more complex and variable, requires sequencing and imqging on shorter timescales, and of course has privacy issues. The European Molecular Biology Laboratory is based across six countries and employs over 1,600 people, but also collaborates with thousands of other scientists and requires access to existing national repositories as well. High-bandwidth networks are therefore necessary to interconnect their on-site computer and storage clusters, but will increasingly be necessary to facilitate connectivity with other research and commercial cloud resources such as EGI.eu and HelixNebula.

David Martin (Argonne National Labs) continued this theme, by presenting on the US Department of Energy’s Exascale Computing Initiative. This aims to develop and operate the next generation of supercomputers at the Argonne, Lawrence Livermore, Los Alamos and Oak Ridge National Labs by 2021, along with a software stack that will present a common computing platform for supporting advanced research applications and neural networks. The Argonne Labs Computing Facility will be based around an Intel Aurora supercomputer with over 1,000 petaflops of processing, 8 PB of memory, and 10 TB/s of input/output capability that will require future network connections in the petabit-per-second range.

Joe Mambretti (Northwestern University) then discussed the Open Science Cloud (OSDC) which is an open-source cloud-based infrastructure that allows scientists to manage, share and analyse large datasets. The aim is to have 1-2 PB of storage at each participating campus, interconnected with 100 Gb/s+ links, but presented and managed as a common namespace with uniform interfaces and policies.

The rest of the day was devoted to how network automation can integrate compute and storage facilities, particularly across multiple domains. Migiel de Vos (SURFnet) presented the work being undertaken for SURFnet 7, and explained the distinction between automation and orchestration whereby the former is considered task and domain specific, whilst the latter is developing intelligent processes that consist of multiple automated tasks across multiple domains. This required the development of new information models, standardised interfaces, automated administration, and then predetermined service delivery agreements.

Gerben van Malenstein (SURFnet) then discussed LHCONE Point-to-Point Service that allowed Layer 2 circuits to be dynamically established between Data Transfer Nodes for exchanging data from the Large Hadron Collider. This was built on the AutoGOLE work which was now enabled on 21 open exchange points. Nevertheless, whilst AutoGOLE was a functional and proven multi-domain system, there was still limited uptake by network services and end-users, which was necessary to completely remove human configuration of network equipment and create a truly global research platform.

Most of the following day was devoted to technical discussions chaired by Lars Fischer (NORDUnet) and Eric Boyd (University of Michigan). These focused around some practical examples of network automation being used at the University of Michigan, a passive network measurement system with programmable querying at 100 Gb/s line rates that was being developed by the IRNC AMIS Project, as well as discussions on how to automate the generation of network topology maps.

Topology maps are useful for users to show how they can reach counterparts in other parts of the world, and where particular services are available. They are also useful as a marketing tool to show investors and stakeholders how they contribute towards creating a truly global infrastructure, and demonstrate how the NREN model is accepted around the world, and for example, the GLIF map has become a somewhat iconic piece of artwork.

Other developments were the establishment of a new exchange point called South Atlantic Crossroads (SAX) based in Fortaleza, Brazil that was expected to interconnect with new cable systems to Angola (SACS) and Portugal (EllaLink), as well as to AMPATH and SouthernLight over the existing MONET connection. There were also plans to build procure a new 100 Gb/s connection from Europe to the Asia-Pacific, from Geneva to Singapore via the Indian Ocean to supplement the existing link from Amsterdam to Tokyo via Russia.

There were further updates on the new KREOnet network which supported 100 Gb/s links between five major Korean cities and Chicago (StarLight) via KRLight, as well as multiple 10 Gb/s links to 11 other Korean cities, Hong Kong and Seattle. The KREOnet-S infrastructure further offered SDN capabilities permitting dynamic and on-demand virtual network slicing, whilst a Science DMZ provided high-performance computing facilities for KISTI’s new 25.5 petaflop supercomputer.

SURFnet is transitioning its network to SURFnet 8 and would be upgrading its core network and international links, whilst StarLight was developing a Trans-Pacific SDN testbed, as well as an SDX for the GENI initiative.

The closing plenary session focused on how high-bandwidth research connections and exchange points can be better planned and coordinated, and whether a new entity should be created to support this. The GLIF Co-Chairs Jim Ghadbane (CANARIE) and David Wilde (AARNet) outlined some ideas around this, and then hosted a discussion on how things should be progressed.

Further Information

Disclaimer: Viewpoints expressed in this post are those of the author and may or may not reflect official Internet Society positions.

Related Posts

Improving Technical Security 15 March 2019

DNS Privacy Frequently Asked Questions (FAQ)

We previously posted about how the DNS does not inherently employ any mechanisms to provide confidentiality for DNS transactions,...

Improving Technical Security 14 March 2019

Introduction to DNS Privacy

Almost every time we use an Internet application, it starts with a DNS (Domain Name System) transaction to map...

Improving Technical Security 13 March 2019

IPv6 Security for IPv4 Engineers

It is often argued that IPv4 practices should be forgotten when deploying IPv6, as after all IPv6 is a...