EGI Conference 2021

Europe/Amsterdam
Zoom

Zoom

Description

Beyond the Horizon: shaping the digital future with advanced computing services

The annual EGI conference will be held - virtually - this autumn, from the 19th until the 21st of October

The final version of the programme is available and the registrations are open! 

The following sessions are organised as co-located events – you can indicate when registering if you want to attend:

 

 

 

Participants
  • Aastha Mathur
  • Abigail McBirnie
  • Ada Pellumbaj
  • Adinarayana Jagarlapudi
  • Adrian Coveney
  • Aggeliki GKAMILIARI
  • Agnes Mika
  • Ahmad Hesam
  • Aida Palacio Hoz
  • Akos Hajnal
  • Alastair Dewhurst
  • Alba Vendrell
  • Alberto Azevedo
  • Alberto P. Marti
  • Alessandra Nuzzo
  • Alessandro Amicone
  • Alessandro Costantini
  • Alessandro D'Anca
  • Alessandro Olivo
  • Alessandro Paolini
  • alex Dehne Garcia
  • Alexander Nakos
  • Alexandr Mikula
  • Alexandra Singer
  • Alexandre Bonvin
  • Alexandru Dinu
  • Alfonso Orta
  • Ali J. Ghandour
  • Aljosa Hafner
  • Alvaro Lopez Garcia
  • Amanda Calatrava
  • Amjad Ullah
  • Amrita Prasad
  • Ana Georgieva
  • Ana Morales
  • Anabela Oliveira
  • Anastas Mishev
  • Anders Tjulin
  • Andre Gemuend
  • Andrea Cristofori
  • andrea giannotta
  • Andrea Manzi
  • Andreas Gavrielides
  • Andreas Kozadinos
  • Andreas Ocklenburg
  • Andreas Peters
  • Andrei Costin
  • Andrei Tsaregorodtsev
  • Andrew Lahiff
  • Anelia Stoikova
  • Angela Lausch
  • Anna Manou
  • Annette Langedijk
  • Annika Kossack
  • Antonio Filograna
  • Antonio S. Cofiño
  • Antonios Zervas
  • Antoon Prins
  • Anwar Vahed
  • Apostolos Theodoridis
  • Ari Asmi
  • Arne Berre
  • Artemis Avgerou
  • Aurelia Andrés
  • Aurélien Leynet
  • Axel Bonnet
  • Baptiste Cecconi
  • Baptiste Grenier
  • Barbara Krasovec
  • Barbara Martelli
  • basuki suhardiman
  • Benjamin Jacobs
  • Benjamin Schumacher
  • Bernd Saurugger
  • Bert Jagers
  • Bjorn Backeberg
  • Bob Jones
  • Bonnie C Carroll
  • Borja Esteban
  • Bouchra Rahim
  • Bratislav Petrovic
  • Brian King
  • bunguli ajalah
  • Carl Simon Adorf
  • Carl-Fredrik Enell
  • Carlos Casorran
  • Carlos Fernandez
  • Carlos Rubia
  • Carmel Walsh
  • Carolin Penke
  • Carolina Odman
  • Catalin Condurache
  • Catarina Ortigão
  • Caterina Barillari
  • Cecilia Cabello
  • Cesare Delle Fratte
  • Charalampos TSITLAKIDIS
  • Charis Chatzikyriakou
  • Chenglai Wu
  • chris de loof
  • Christian Briese
  • Christian CUCINIELLO
  • Christian Pagé
  • Christoph Witzig
  • Christophe Laures
  • CHRISTOS PAPACHRISTOS
  • Chuck Cook
  • Claudia Cesarini
  • Claudia Martens
  • Claudio Pisa
  • Clementina Piani
  • Connor Pettitt
  • Corneliu Caileanu
  • Cyril Lorphelin
  • Cécile Cavet
  • Dai Sato
  • DANIEL FODOREAN
  • Daniel Kouril
  • Daniel Volovici
  • Daniele Spiga
  • David Aikema
  • David Groep
  • David Kelsey
  • David Schwartz
  • David Southwick
  • Davide De Santis
  • Davide Salomoni
  • Dejan Lesjak
  • Denis Pugnere
  • Di Wu
  • Diana Gudu
  • Dick Schaap
  • Diego Scardaci
  • Dimple Sokartara
  • Dmytro Dehtyarov
  • Dobriša Dobrenić
  • Doina Cristina Duma
  • Donatello Elia
  • Doru BALAN
  • Dragan Radolović
  • Dragoș-Cătălin Barbu
  • Drahomira Cupar
  • Eduard Jacob
  • Eduard Jacob
  • Eduardas Kutka
  • Effie Zafeirakopoulou
  • Ela Daci
  • Elena Gazzarrini
  • Eleni Petra
  • Eleonora Testa
  • Elisa Cauhe
  • Elke Dall
  • Elli Papadopoulou
  • Elli Papadopoulou
  • Emir Imamagic
  • Emmanouil Vamvakopoulos
  • Enol Fernandez
  • Enrico Vianello
  • Eric Yen
  • Esa Keskitalo
  • Esa-Pekka Keskitalo
  • Eugene Krissinel
  • Eva Pastorková
  • Exams Hero
  • Fabio Hernandez
  • Fabio Perossini
  • Fabrizio Antonio
  • Fabrizio Furano
  • FEDERICO CALZOLARI
  • Federico Drago
  • Fernando Aguilar
  • Fernando Rodriguez
  • Feyza Eryol
  • Floris Hermanns
  • Fotis Karagiannis
  • Francesco Tusa
  • Frank Wuerthwein
  • Fredrik Lindén
  • gabriel marcel ikopi moleko
  • Gabriel Neagu
  • Gabriel Zachmann
  • Geneviève Romier
  • George Papastefanatos
  • GEORGIOS KAKALETRIS
  • Gerard Coen
  • Germán Moltó
  • Giacinto Donvito
  • Giacomo Inches
  • Gianni Dalla Torre
  • Gino Marchetti
  • Giovanni Rimassa
  • Giuseppe La Rocca
  • Giuseppe Lo Presti
  • Giuseppe Maulucci
  • Gonzalo Merino
  • Greg Corbett
  • Guenter Grein
  • Guido Aben
  • Gwen Franck
  • Gyula Mester
  • Gábor Hermann
  • Hadia Yaseen
  • Hakan Bayindir
  • Hamed Hamzeh
  • HANDE ERDEM
  • Happy Sithole
  • Helen Clare
  • Helmut Sverenyak
  • Hermann Heßling
  • Horst Schwichtenberg
  • Hrachya Astsatryan
  • Humaira Abdul Salam
  • Hussein Sherief
  • Ian Collier
  • Ignacio Blanquer Espert
  • Ignacio Lamata Martinez
  • Ignacio Peluaga Lozada
  • ii ii
  • ilirjan malollari
  • Indika Priyantha Kumara Dewage
  • Ingemar Haggstrom
  • IOANNIS LEDAKIS
  • Irena Vipavc Brvar
  • Irene Sánchez
  • Iryna Kuchma
  • Isaac Motochi
  • Isabel Campos
  • Istvan Pintye
  • Iulia Popescu
  • Iván Analuisa
  • Iván Díaz Álvarez
  • Izabela Zrazinska
  • Jan Astalos
  • Jan Studnicka
  • Jan Wiebelitz
  • Jana Broncova
  • Jana Misic
  • Jason Brudvik
  • Jay Greenfield
  • Jayesh Wagh
  • Jean Baptiste Fankam Fankam
  • Jens Jensen
  • Jeonghoon LEE
  • Jeremie Lengrais
  • Jeremy Azzopardi
  • Jeremy Smith
  • Jerry Horgan
  • Jesús Gorroñogoitia
  • jianhui li
  • Jiri Chudoba
  • Jiri Sitera
  • Jo Morris
  • Joao Costa
  • Joao Fernandes
  • Joao Pina
  • John Helliwell
  • John Ronan
  • Jonathan Tedds
  • Jorge Gomes
  • JOSE CABALLERO BEJAR
  • Jose Flix
  • Jose-Maria Carazo
  • Joseba Sanmartín
  • Joseph Wafula
  • José Antonio Díaz
  • José María García
  • jounaid ruhomaun
  • Juha Herrala
  • Julie Arteza
  • Junyi Wang
  • Jurica Spoljar
  • Jurry de la Mar
  • Justin Clark-Casey
  • Jérôme Pansanel
  • József Kovács
  • Kaori Otsu
  • Karolj Skala
  • Katharina May
  • Kay Graf
  • Kechil Kirkham
  • Kilian Schwarz
  • Klaas Wierenga
  • Konstantinos Kagkelidis
  • Konstantinos Oikonomou
  • Kostas Koumantaros
  • Kostas Moutselos
  • Kristiina Himanen
  • Kristýna Lyžbická
  • Ladislav Hluchy
  • Larisa Ungurean
  • Laura del Cano
  • Laura Gavrilut
  • Laure Barbot
  • Lauren Michael
  • Laurent Caillat-Vallet
  • Lene Krøl Andersen
  • Leslie Groer
  • Levente Farkas
  • Lianglin Hu
  • Lieven RAES
  • Lili Zhang
  • Liliana Gherghe
  • Linda Cornwall
  • Lisana Berberi
  • Liza Ocklenburg
  • Lottie PROVOST
  • Louise Bezuidenhout
  • Luca Bolognini
  • Luciano Gaido
  • Ludovica Paseri
  • Luigi Colucci
  • Lukas Burgey
  • Luuk Uljee
  • Lyubka Pashova
  • Maarten Kremers
  • Maarten Litmaath
  • Maciej Pawlik
  • Malgorzata Krakowian
  • Mandy Yuju Lin
  • Manuel Parra
  • Marcin Plociennik
  • Marcin Sieprawski
  • Marco Rorro
  • Marco Verlato
  • Marcus Bahn
  • Marcus Hardt
  • Maria Girone
  • Maria Lindqvist
  • Maria McCann
  • Maria Stoica
  • Marica Antonacci
  • Marie-Louise Weijde
  • Marieke Willems
  • Marina Grubišić
  • Mario Lassnig
  • Mario Reale
  • Marion Pause
  • Marius Appel
  • Mark Dietrich
  • Marta Gutierrez David
  • Marta Rodrigues
  • Martin Atanasov
  • Martin Belavic
  • Martin Bobak
  • Martin Gutbrod
  • Matthias Schramm
  • Matti Heikkurinen
  • Mehmet Satoglu
  • meiyu hao
  • Melina Jander
  • Melissa Cragin
  • Metodiyka Tarlyovska
  • Michael Kuss
  • Michael Schuh
  • Michaela Bežová
  • Michela Bertolotto
  • Michela Milano
  • Michele De Bonis
  • Michelle Barker
  • Miguel Caballer
  • Mihai Patrascoiu
  • Mike Jacksteit
  • Mikko Grönroos
  • Milan Ojsteršek
  • MIRCEA BULINSKI
  • Mirek Ruda
  • Miriam Saso
  • MIron Livny
  • Miroslav Dobrucky
  • Mirvat Al jogami
  • Mischa Sallé
  • Mohammed Humran
  • Monica Balatresi
  • Muhammad Aleem Sarwar
  • Muhammad Hilmy
  • Nadica Curic-Segaric
  • Nadine Locatelli
  • Natalia Manola
  • Natalie Danezi
  • Nick Choi
  • Nicolas Liampotis
  • Nicoletta Carboni
  • Nikos Chondros
  • Norbert Schmidt
  • Noshaba Cheema
  • Nurzatil Mat Jalaluddin
  • Olga Rodikow
  • Oliver Rack
  • Olivier Rouchon
  • Owen Appleton
  • Pablo Fernandez
  • Pablo Sánchez Núñez
  • Panayiotis Michael
  • Paola Materia
  • Paolo Manghi
  • Pascal Derycke
  • Pascal Heus
  • Pasquale Pagano
  • Patricia Herterich
  • Patrick Fuhrmann
  • Patrick Griffiths
  • Patryk Lasoń
  • Paul Coddington
  • Paul Millar
  • Paul Musset
  • Paul Provakar Saha
  • Pauli Assinen
  • Pavel Weber
  • Peter Bogatencov
  • Peter Kacsuk
  • Peter Selsam
  • Peter Szegedi
  • Petr Pospíšil
  • Phillip Bodare
  • Pierre-Etienne Macchi
  • Pinja Koskinen
  • pnassisi pnassisi
  • Povilas Treigys
  • Radu Mutihac
  • Ralph Thesen
  • Ramaz Kvatadze
  • Raymond Oonk
  • Renata Słota
  • Renato Santana
  • Resmi Arjun
  • Ricard Munné
  • Ricardo Martins
  • Riccardo Bianchi
  • Riccardo DI MARIA
  • Richard Dennis
  • rita sofia meneses
  • Rob Simmonds
  • Robert Hanisch
  • Robert Huber
  • Robert Lovas
  • Roberto Pasciuti
  • Roberto Sabatino
  • Roderick Mooi
  • Rodrigo Augusto Dias
  • Rodrigo Honorato
  • Rolandas Naujikas
  • Romain David
  • Romain Wartel
  • Ron Dekker
  • Ron Trompert
  • Rose Cooper
  • Rowan Arts
  • Rowan Brownlee
  • Rui Figueira
  • Rémi Cailletaud
  • Sa-kwang Song
  • Sabine Crepe-Renaudin
  • Sagar Dolas
  • Salima Rehemtula
  • Samuel Bernardo
  • Sandro Fiore
  • Sebastian Lopienski
  • Sebastian Luna-Valero
  • Sebastián Risco
  • Selcuk Bilmis
  • Sergio Andreozzi
  • Sergiy Svistunov
  • Shaun de Witt
  • Sheng-Yuan Liu
  • Shreyasvi Natraj
  • Shubo Chakrabarti
  • Siamak Farshidi
  • Sid El Moctar Ahmed Maouloud
  • Silvio Pardi
  • Simon Hodson
  • Simona Coman
  • Simona Hudecova
  • Slavek Licehammer
  • Slavica Jonic
  • Smitesh Jain
  • Sofía Pérez
  • Sonia Santiago
  • Sophie Servan
  • Sorina POP
  • Sorina Ropciuc
  • Spiros Koulouzis
  • Sridhar Gutam
  • Stefano Bagnasco
  • Stephan Kindermann
  • Stephan Siemen
  • Susana Sánchez Expósito
  • Sven Gabriel
  • Sy Holsinger
  • Sylvie JOUSSAUME
  • Tamas Kiss
  • Teresa Gomez-Diaz
  • Themis Zamani
  • Thomas Dack
  • Thomas Eifert
  • Tiago Gonçalves
  • Tibor Kálmán
  • Tibor Molnar
  • Tigran Mkrtchyan
  • Tim Noble
  • Tim Wetzel
  • Tiziana Ferrari
  • Tobias Kerzenmacher
  • Tomas Recio
  • Tommaso Boccali
  • Tomoaki Nakamura
  • Torill Hamre
  • Uros Stevanovic
  • Valentin Kozlov
  • Valentino Cavalli
  • Valeria Ardizzone
  • Vanessa Bremerich
  • Vanessa Hamar
  • Vas Vasiliadis
  • Vasso Kalaitzi
  • Verena Wottrich
  • Vicky, Pei-Hua HUANG
  • Viet Tran
  • Ville Tenhunen
  • Vincent Negre Vincent
  • Vincenzo Spinoso
  • Vitalie Cervinschi
  • Volker Beckmann
  • Volker Guelzow
  • William Karageorgos
  • Xavier Espinal
  • Yannick LEGRÉ
  • Yin Chen
  • Yona Baskharoun
  • Yosef Moatti
  • Yuan CHAO
  • Yuri Demchenko
  • Zachary Smith
  • Zdenek Sustr
  • Zhengdong SHI
  • Zhiming Zhao
  • Zoe Vasileiou
  • Zoltán Farkas
  • Zoltán Gál
  • Zoran Bekić
  • Štefica Dodig
    • 1
      EGI101 - Introduction to EGI go.egi.eu/egi2021-101

      go.egi.eu/egi2021-101

      Zoom

      Are you new to EGI? And maybe a bit confused about what EGI is, what we do, how we are structured and how you can collaborate with us? Then this is the session for you!

      The EGI Federation is an international e-Infrastructure to provide advanced computing and data analytics services for research and innovation. EGI federates compute resources, data management services, support teams and various online, thematic services from over 30 countries and international institutes, to make those available for researchers, innovators and educators. Since 2021 EGI delivers a compute platform in the European Open Science Cloud through the H2020 EGI-ACE project. This session will provide a basic introduction to EGI, covering all the fundamental topics that you need to know before deep-diving into the EGI'21 conference. The session will include a presentation and a Q&A slot.

      Speaker: Gergely Sipos (EGI.eu)
    • 2
      Opening and Welcome go.egi.eu/egi2021-1 (Zoom Room 1)

      go.egi.eu/egi2021-1

      Zoom Room 1

      Speaker: Arjen van Rijn (NIKHEF)
    • 3
      Keynote: Scientific Computing with EGI: Current Status and Future Outlook go.egi.eu/egi2021-1 (Zoom Room 1)

      go.egi.eu/egi2021-1

      Zoom Room 1

      What is the impact of EGI on science and how are researchers' needs transforming scientific computing solutions in the coming decade? The EGI Foundation Director, Dr. Tiziana Ferrari will present the latest status of the EGI Federation and its contribution to EOSC focusing on its scientific impact.

      Speaker: Tiziana Ferrari (EGI.eu)
    • 4
      Keynote: 50 years of IN2P3 - support for scientific ground breaking discoveries and future perspectives (Pierre-Etienne Macchi, Director IN2P3 Computing Center) go.egi.eu/egi2021-1 (Zoom Room 1)

      go.egi.eu/egi2021-1

      Zoom Room 1

      Former EGI.eu Executive Board member, Pierre-Etienne Macchi is the director of the IN2P3 Computing Center, a thematic computing center dedicated to research in high energy physics, nuclear physics and astroparticles.
      He acts as the IN2P3 representative at the CNRS e-infrastructures committee and he participates, in close coordination with the IN2P3 management, to the policy definition of the Institute regarding computing. He attends the Group Council of France Grilles, to which IN2P3 is the main resource provider. He is also the CNRS representative at the French NREN (RENATER) board of users, and participates in various reflection groups set up by the French research ministry about data centers and research data organization.

      Speakers: Arjen van Rijn (NIKHEF), Pierre-Etienne Macchi (CNRS-IN2P3)
    • 5
      Keynote: 50 years of SRCE - achievements and future perspectives for scientific computing of SRCE (Ivan Marić, SRCE) go.egi.eu/egi2021-1 (Zoom Room 1)

      go.egi.eu/egi2021-1

      Zoom Room 1

      Ivan Marić is the Director of University Computing Centre - SRCE, where he began his career in 1990. During his time in SRCE, he played an important role in the creation and
      construction of the Croatian academic and research network at the beginning of 1990’s and has since then participated in numerous national initiatives related to building Croatian e-infrastructure. Currently, he leads the national strategic project Croatian Scientific and Educational Cloud (HR-ZOO), which aims to build a new generation of e-infrastructure for advanced science and modern education.

      On the international level, he is actively involved in building and governing major pan-European e-infrastructures like GEANT and EGI. Mr. Maric holds several positions in
      international organizations as national representative for Croatia - he is an official Croatian delegate in the EOSC Steering Board and in EOSC Association General Assembly. He is the Croatian representative in the Future Internet Forum (FIF) from 2016, and an e-IRG delegate from 2006.

      He is also the Croatian representative in Horizon Europe Research Infrastructure Programme Committee.

      Speakers: Arjen van Rijn (NIKHEF), Mr Ivan Maric (Director of University Computing Centre - SRCE)
    • 11:00 AM
      Coffee break
    • Collaboration Accross Boundaries - Presentations go.egi.eu/egi2021-2 (Zoom Room 2)

      go.egi.eu/egi2021-2

      Zoom Room 2

      Convener: Tamas Kiss (University of Westminster, London, UK)
      • 6
        Coping with growing demands by cooperation across universities on the IT service level

        According to our observation, universities‘ IT service departments either get bogged down by the multiplicity of (valid) demands or simply is unable to fulfil a portion of these demands. On the other hand, users often do not even know about those services that exist. We learned that from a short survey conducted as part of a project. So, IT has to cope with the set union of all users’ demands. In the meantime, the same users, due to this lack of information, add new demands. Often, IT‘s standard way to cope with this situation is to confine to “standardized“ services. However, in the days of various cloud service offerings – public or private – such a confinement bears the risk that these standardized services compete directly with services available from the cloud. In the meanwhile, the individualization that has been avoided is now deeply missing since it could be the value to the academic users that IT could add to cloud services.
        Our presentation shows examples of this and other recent user surveys that reveal parts of the expectations from IT, and real-world examples of cross-university collaborations infrastructure that allow to leverage the positive scaling effects of bundling and simultaneously create the freedom to engage in individualized service flavors.
        The first example is the creation of a federated identity management. While cross-site or even global authentication is well known for stateless services like access to WiFi by eduroam, for true collaborative services we need persistent authentication that allows to tie group memberships and roles to an identifier. To our knowledge, the handling of groups – based on accounts or on identities - is still most often done within each particular IT service. This is perfect in the sense that the interpretation of a certain role depends on the particular service when it comes to granting special permissions etc. However, the definition of a project group has the same meaning, independently of a particular service. From our perspective an identity based, service-spanning authentication and authorization is inevitable either to build collaborating in IT systems as well as to run cross-site workflows that imply any sort of distinct roles. This the more since from a scientific perspective a separate view on resources and services like computing, data storage, collaboration etc. is outdated. Thus, the creation of a process infrastructure for the easy use of cross-university services is an essential building block that subsequently enables integration into higher-level infrastructures.
        Our “example service” is the newly created cross-university backup-service in a part of Germany, covering 42 universities. The support model as well as the role concept that is necessary to establish a tiered self-service architecture relies on federated identity management mentioned above.
        We state that these examples show how IT can cope with current challenges to deliver services of high quality and make them accessible to a wide range of users. Simultaneously, by collaborative service delivery, IT regains free room to innovate services and to enhance the “standard” services by individual, user specific add-ons.

        Speaker: Thomas Eifert (RWTH Aachen University)
      • 7
        Identity management beyond a simple VO management

        Federated authentication and authorization management represented by AARC Blueprint Architecture is naturally stressing out the importance of the part represented by the proxy component. Indeed, the proxy component has an essential role because it handles the connection to the home IdP or other authentication sources and processes all the attributes, which are consequently passed downstream.

        Another crucial part of AARC Blueprint Architecture is an identity management system (IdM). Its role is sometimes perceived only as a tool for managing virtual organizations (VO) and groups. That is a logical first step when building new authentication and authorization infrastructure (AAI). Having the option to enrol new users and manage them in VOs and groups is often enough to support basic use-cases for AAI. But in reality, the role of IdM is much broader and utilizing its whole range enables pushing capabilities of AAI to an entirely new level.

        The role of IdM is to provide storage and a centralized view of all identity-related data. It can also ensure that relevant data is provisioned (and deprovisioned) to any services that need them. Proxy is only one from many components that are governed by the IdM system. There are other systems and services where user accounts have to be created and maintained even though they do not access them using federated authentication through proxy. Good examples are directories services or mailing lists.

        IdM is not only about provisioning but also about importing data from other sources. Therefore the IdM system can be used as a central component that gathers data from all relevant authorities, including data self-provided by users (e.g. during registration), processes them, applies policies, and then is able to provision them to targets. Furthermore, it maintains consistency among all the connected systems during the whole user-life cycle.

        Except for the automated processes described before, IdM enables managing attributes, groups, roles and other entitlements. People having the manager role can decide who will be authorized to which service and configure other attributes to fine-tune authorization rules and properties of managed accounts. Then the automatic process of IdM takes over and makes sure the configuration is provisioned to proper targets.

        This presentation will demonstrate a wide range of IdM features using the IdM system Perun, which is quite a known tool for federated identity management. The features will be explained on real-world use-cases gathered from existing instances of Perun and communities using it, like EGI, ELIXIR, BBMRI-ERIC or Czech national e-Infrastructure.

        Speaker info: https://www.muni.cz/en/people/255920-slavek-licehammer

        Speaker: Slavek Licehammer (CESNET)
      • 8
        Implementation of FAIR principles in nuclear magnetic resonance (NMR) spectroscopy experimental workflows

        Over the past decades, the scientific domain has seen an exponential growth of data generation. As expected, the storage, management and analysis requirements are growing proportionally, with a direct impact on the scientific output. Certain scientific communities like the astronomical one, have already organised and standardised computational workflows, whereas others can benefit from modern good practice principles. These are called FAIR principles and are currently actively supported by the ecosystem of the European Open Science Cloud (EOSC).

        This contribution presents the ongoing activities regarding the implementation of FAIR principles in a Nuclear Magnetic Resonance (NMR) spectroscopy experimental workflow, performed at the CERIC-ERIC partner facility, NMR centre at the National Institute of Chemistry (Slovenia). The experiments carried out at this facility range from biological and pharmaceutical studies to studies in energy and basic matter research of high impact. The complete data life cycle has been upgraded, including new approaches to data and metadata acquisition, storage, archiving, processing and visualisation through web and cloud services, paving the way towards advanced data science on the measured datasets. Furthermore, integration of the entire pipeline into the open-source Jupyter web services is also in development. The goal is fostering cross-discipline open science while improving efficiency and accessibility. This is done in the context of a major EU project (PaNOSC) that connects six European large-scale research centres and two e-infrastructures, providing photon- and neutron-based analytical techniques.

        The reported results outline certain challenges encountered during the conceptualization and implementation of the NMR workflow upgrade, e. g. handling experimental data from offline devices. At the same time, the aim for the results is to be useful to other scientific fields in the process of adopting FAIR principles and more efficient data handling, tracking and processing.

        Speaker bio: https://www.linkedin.com/in/aljosa-hafner-99350b174/

        Speaker: Aljosa Hafner (CERIC-ERIC - Central European Research Infrastructure Consortium)
      • 9
        Overview of incidents we are dealing with and how to improve the response readiness.

        So far year 2021 has shown to the EGI CSIRT team that the services for
        research and science are not immune to the dark side of the web.
        Compromised platforms can be used to malicious activities and EGI
        infrastructure surely has interesting targets for potential abuse. It is
        ever more obvious that information security must be constantly cared
        for. More complex layered services spanning to multiple vendors,
        platforms and services bring additional challenges to deployment and
        operations, which require our special attention

        Speaker: Pinja Koskinen
      • 10
        Incident Response in multiple Service-User-Relation layers and threat intel sharing in challenging environments.

        Incident Response in complex governance environments requires an additional level of responsibility awareness and clear agreements on the roles in Incident Response.
        We will look at two scenarios to illustrate the problem.
        1. Multiple layers of Service-User-Relations in cloud environments. Here we have the Resource Center providing the virtualization enabling technologies as a service, the users can then run another IAAS on the VM provided by the User, who then in turn provides a service to his users,
        that can run services for another layer of users. To keep track of who is responsible for what and has sufficient information to act on reported problems is getting hazy the more such layers exist.
        A similar problem exist in environments like a federation of
        federations, which may have different policies and procedures in effect. Information sharing, in particular Threat Intel sharing in these set ups are an additional challenge, possible solutions will for this be presented here.

        Speaker bio: https://nl.linkedin.com/in/sven-gabriel-a3296227

        Speaker: Sven Gabriel (NIKHEF)
      • 12:30 PM
        Q&A
    • EOSC - Workshop: Bridging from theory to practice with EOSC-Synergy go.egi.eu/egi2021-3 (Zoom Room 3)

      go.egi.eu/egi2021-3

      Zoom Room 3

      The project EOSC-Synergy is pushing the state-of-the-art to facilitate the adoption of EOSC for infrastructure managers and scientific application developers in a way that supports EOSC Core interoperability guidelines, fosters the development of quality software, and provides a comprehensive environment for Thematic Services to become a part of the EOSC Exchange ecosystem.

      Our strategy to foster EOSC adoption is based on the development of a quality-based approach built upon the definition of a software and service quality baseline, that enables the safe deployment of scientific thematic services, in terms of security and safe data transfer across Europe. Following the EOSC core architecture, EOSC-Synergy has integrated national e-infrastructures and developed a portfolio of services to support 10 thematic services to support relevant scientific areas in Europe.

      The purpose of this workshop is to share our vision on all the project developments from a pragmatic point of view, in the form of hands-on practical approach. The target audience includes service and infrastructure providers, research applications developers, and support software developers.

      The sessions will cover the several steps of the life cycle, starting by the infrastructure approach, with a focus on federated access to such infrastructures such as Cloud and HPC mainframes. Here the project will present the developments, “HowTos”, and demonstrations that streamline access to Clouds and HPC resources.

      On the software quality side we will present the status of our quality assurance developments for software and services, organized around the implementation of the SQA as a Service (SQAaaS) platform, and data FAIRness, and how they fit in the framework of the EOSC Interest Group on “Infrastructure for Quality Research Software”. This will be followed by a practical tutorial on how to use the SQAaaS platform to integrate automated quality assurance procedures in the development of research software.

      We will also portray a selection of thematic services that have been integrated among our 10 pilot cases. On the use cases and applications side, we will count as well with invited presentations from cooperating projects to analyse the synergies and opportunities for cooperation aiming at integrating thematic services in EOSC.

      The training portal of EOSC-Synergy will be used as a tool during the session to showcase the best practices approach to the development of training material, and collect feedback from the participants.

      Conveners: Isabel Campos (CSIC), Valentino Cavalli (EGI.eu)
      • 11
        Introduction to the workshop
        Speaker: Isabel Campos (CSIC)
      • 12
        Increasing Quality and Capacity of Thematic Services in EOSC-SYNERGY

        "Ignacio Blanquer is full professor in the Computer Systems department. Since 2015, he leads the Grid and High Performance Computing Research group and since 2019 he is the vice-director of the Institute of Instrumentation for Molecular Imaging. Ignacio Blanquer has participated in over 60 European and National projects, and he has been the coordinator of 3 European Projects in the area of cloud computing and data analysis. Ignacio Blanquer is the coordinator of the Spanish Network for e-Science and he serves as an expert to the General Secretary of Research in the Spanish Ministry of Science and Innovation.
        He is the author and co-author of nearly 55 articles in indexed journals and over 100 publications in book chapters, non-indexed journals and international congress proceedings."

        Speaker: Ignacio Blanquer (UPVLC)
      • 13
        Service Quality Assurance as a Service: Fostering Service Integration in EOSC

        Speaker bio: https://www.lip.pt/?section=about&page=person-details&details&id=839

        Speaker: Samuel Bernardo (LIP)
      • 14
        SQAaaS Web Demo: so simple that you can’t miss it
      • 15
        Invited presentation and discussion: EXPANDS use cases and quality approach
        Speaker: Patrick Fuhrmann (DESY)
      • 16
        Best practices in tutorials development: the EOSC-Synergy learning portal
        Speaker: Helen Clare (JISC)
    • EOSC Workshop: EGI-powered Thematic Services in EOSC go.egi.eu/egi2021-4 (Zoom Room 4)

      go.egi.eu/egi2021-4

      Zoom Room 4

      The EGI federation underpins many of the services of EOSC in different ways: providing compute and storage for hosting and scale-out; Enabling access via user authentication-authorisation; Monitoring service availability and reliability; Interfacing with Helpdesk, Marketplace and order management.
      During the EOSC-hub project (2018-2021) the EGI federation worked with over 30 scientific community providers and empowered them in EOSC in these various ways, resulting in over 50 thematic services that are now accessible for open science via the EOSC Marketplace. The same route is followed since January in the EGI-ACE and EOSC Future projects, where over 20 communities are supported so far to make most of the EGI federation, to reach the functionality and maturity EOSC users require.

      This workshop would introduce the most prominent support mechanisms that the EGI federation offers for EOSC Thematic Services, via real use cases that would be presented by scientific communities (providers of Thematic Services). The presentations would demonstrate the diversity of support mechanisms that EGI offers across technologies, countries and scientific disciplines. The session would be a showcase of some of the Thematic Services that were achieved in EOSC-hub, supported in EGI-ACE or EOSC Future. The presentations would show the impact of this work, ultimately demonstrating the impact of EGI on Open Science.

      The session would be relevant for

      • scientific end users who want to see some of the most popular Thematic Services that are offered in EOSC.
      • service developers and providers to see examples of how EGI could empower and bring their setup into EOSC.
      • compute and data providers to see values in joining EGI and EOSC.
      Convener: Giuseppe La Rocca (EGI.eu)
      • 17
        Welcome

        Introduce the methodology used by EGI to support Thematic Services (with technical services and w. Competence Centres)

        Speaker: Giuseppe La Rocca (EGI.eu)
        • a) Welcome
      • 18
        OPENCoastS+ - moving from on-demand coastal circulation to water quality forecasts
        Speaker: Anabela Oliveira (National Laboratory for Civil Engineers)
      • 19
        How AiiDAlab makes complex scientific workflows accessible and shareable
        Speaker: Dr Carl Simon Adorf (EPFL)
      • 20
        Status of the LOFAR radio-astronomy Thematic service

        Speaker bio: https://www.linkedin.com/in/agnesmika/

        Speaker: Agnes Mika (ASTRON)
      • 21
        WeNMR under EGI: A 10+ years happy symbiosis
        Speaker: Alexandre Bonvin (eNMR/WeNMR (via Dutch NGI))
      • 22
        Closing and wrap-up
        Speaker: Giuseppe La Rocca (EGI.eu)
    • Envisioning the Future - Presentations go.egi.eu/egi2021-1 (Zoom Room 1)

      go.egi.eu/egi2021-1

      Zoom Room 1

      In this session, there are three different contributions:

      • Dr. Yuri Demchenko will present a vision for RI platforms as a Service that incorporates recent digital technologies and platform/ecosystem thinking; this work is part of the project SLIDES-DS and represents a design study towards the future SLICES Research Infrastructure part of the ESFRI Roadmap

      • Dr. Fotis Karagiannis will share insights from the upcoming e-IRG white paper entitled "Good practices of coordination within and across e-Infrastructures and thematic Research Infrastructures (RIs)" explaining how the different levels of federation are key to building international infrastructures.

      • Ms. Ludovica Paseri will address the topic of open science from the legal perspective, starting with the fundamental rights on which it is rooted.

      Convener: Sergio Andreozzi (EGI.eu)
      • 11:15 AM
        Introduction
      • 23
        Future Scientific Data Infrastructure: Towards Platform Research Infrastructure as a Service (PRIaaS)

        Modern Science is becoming increasingly data driven and works with a large amount of data, which are heterogeneous, distributed and require special infrastructure for data collection, storage, processing, and visualisation. Science digitalization, likewise industry digitalization, is facilitated by the explosive development of digital technologies and cloud based infrastructure technologies and services.

        Current development and establishment of the European Open Science Cloud (EOSC) provided a strong basis for creating federated data infrastructure for European research and bringing to practice the FAIR (Findable, Accessible, Interoperable, Reusable) data principles of the Open Science. EOSC allow for effective data exchange and integration across scientific domains, making scientific data a valuable resource and a growth factor for the whole digital economy and society. To uncover the potential of the future digital and data driven science, the future Scientific Data Infrastructure (SDI) must provide a platform for effective use of scientific data by providing functionality to automate creating specialized/customised ecosystems supporting full cycle of the value creation from data collection to model creation and knowledge acquisition and exchange. Shift of the focus from infrastructure operation to value creation will require new FutureSDI design approach, operation and evolution to respond to changing requirements and evolving technologies. Growing infrastructure complexity will require automation of the infrastructure provisioning and operation, allowing researchers to focus on problem solving. Responsibility and sustainability principles must be applied at all stages of the created instant infrastructure from the design to operation, monitoring and key performance indicators must be assessed continuously, presumably assisted by AI optimisation algorithms.

        This paper presents two lines of analysis: one is retrospective analysis related to the European Research Infrastructure (RI) development stages and timeline from centralized to distributed and current Federated Interoperable; another storyline provides analysis of digital technologies trends and identifies what technologies will impact the future SDI.

        Based on this analysis, the paper proposes a vision for the future RI Platform as a Service (PRIaaS) that incorporates recent digital technologies and enables platform and ecosystem model for future science. Notably the proposed PRIaaS adopts TMForum Digital Platform Reference Architecture (DPRA) that will simplify building and federating domain specific RIs while focusing on the domain specific data value chain with data protection and policy based management by design.
        This research is a part of the SLICES-DS project that represents a design Study stage for the future SLICES Research Infrastructure for digital technologies that is a part of the ESFRI Roadmap.

        About the speaker:
        Yuri Demchenko is a Senior Researcher at the Complex Cyber Infrastructure Research Group of the University of Amsterdam. He is graduated from the National Technical University of Ukraine "Kiev Polytechnic Institute" where he also received his PhD degree. His main research areas include Data Science and Data Management, Big Data and Infrastructure and Technologies for Data Analytics and Artificial Intelligence, DevOps and cloud based software development, general security architectures and distributed access control infrastructure. He is currently involved in the European projects GEANT4, SLICES-DS where he develops different elements of cloud based infrastructures for scientific research and trusted data sharing, as well as projects MATES, FAIRsFAIR that address aspects of skills management and organizational capacity building.

        Speaker: Dr Yuri Demchenko (University of Amsterdam)
      • 24
        White paper: Good practices of coordination within and across e-Infrastructures and thematic Research Infrastructures (RIs)

        The presentation will provide insights on the first iteration of new e-IRG White Paper 2021 that is entitled "Good practices of coordination within and across e-Infrastructures and thematic Research Infrastructures (RIs)", while the overall White Paper (with more iterations planned later in 2021) is entitled "Vision for an inclusive and holistic e‑Infrastructure ecosystem for the European Research Area".

        Following a series of efforts on national coordination (e-IRG Roadmap 2016, e-IRG National Nodes paper in 2019), the e-IRG has worked further on coordination practices at institutional, national and regional (cross-country) levels.

        Coordination paradigms and practices are presented around Europe within e-Infrastructures, across e-Infrastructures and thematic RIs and on open science, covering aspects like ownership/governance, organisation (top down/bottom up, distributed/centralised), funding/cost sharing, resource sharing/access policies and the impact of national RI roadmaps in the country organisation.

        Hybrid models of incentivising both e-Infrastructures and thematic RIs are presented, so that the latter requirements (eNeeds) are well captured by the e-Infrastructures. The first iteration is expected to be finalised in October 2021. The presentation will be made by a relevant e-IRG Working Group member. This document is relevant to the EOSC Secretariat study on NOSCIs, although e-IRG remit is broader covering also other areas such as networking and HPC.

        Impact: European structures such as EOSC are based on a chain of institutional/national/regional structures. The paper highlights this federation chain and the importance of institutions being well represented by their national structures at EU level. The e-IRG paper is thus relevant to NOSCIs. Although no one size fits all, the different paradigms presented and the analysis performed can inspire other countries to use models that may be relevant for them and help NOSCs to be better prepared for EOSC.

        About the speaker
        Dr. Fotis Karayannis is an international vendor with more than 25 years of experience in the Information and Communication Technologies domain, focusing mainly on research infrastructures, and in particular on networking, computing and data electronic infrastructures.
        He received his PhD in 1998 in the fields of Integrated Communications and Management of Broadband Networks from National Technical University of Athens, Greece. He has worked for major commercial or research entities such as OTEPlus (ex-incumbent telecom operator consulting company), the Greek Research and Technology Network GRNET (GR) for 8 years, the IT department of the European Organisation for Nuclear Research CERN (CH) for 2 years, the Czech Research Network CESNET (CZ) for 2 years, Trust-IT Services for 2 years, Microsoft Research Cambridge (UK) and Microsoft Innovation Center Greece for 8 years, The Netherlands Organisation for Scientific Research - NWO (NL) for 5 years, ATHENA Research Center for 10 years, Athens University of Economics and Business (GR) for 3 years and the UK Research and Innovation/Science and Technology Facilities Council (UK) for 8 years.
        He has worked for major research/electronic infrastructure initiatives such as the Pan-European Research Network GEANT, The Advanced Computing Infrastructure for Research EGI, the PRACE Supercomputing Research Infrastructure, the e-Infrastructure policy body e-Infrastructure Reflection Group (e-IRG), both as a delegate (5 years) and a member of its secretariat (15 years), the Support to the Reinforcement of the European Strategy Forum For Research Infrastructures (StR-ESFRI and StR-ESFRI2) and the Research Data Alliance Europe (RDA-Europe) project series (RDA Europe-RDA-Europe 4). He served as the coordinator of the e-FISCAL project on estimating computing costs of major e-Infrastructures and also of the ICRI2014 project, responsible for organising the International Conference on Research Infrastructures in 2014. He has also acted as an expert and deputy national delegate in the FP7 Program Committee on Research Infrastructures (2 years). He endeavoured entrepreneurial activities in the area of big data analytics and cloud computing brokerage as the founder and director of Uranus Computing Limited and shareholder of Innov-Acts Limited.
        He authored the document "A Marketplace for e-Infrastructure services", which constituted the basis for the definition of the European Open Science Cloud (EOSC). He also acted as a member of the evaluation committee for the assessment of the EOSC first phase in June 2020.

        Speaker: Fotis Karagiannis (Independent)
      • 25
        Envisioning the future of Open Science by focusing on fundamental rights

        In the current era of the digital revolution, the impact of new technologies and ICTs on scientific research is clearly significant. The radical transformation that the use of technology has determined in science urges a rethinking of the way science is done: the emergence of the Open Science paradigm is a step in this direction. Open Science does not only aim at promoting the openness of science: the goal is to support good science, in a broad sense, based on communication, cooperation, minimising waste of effort, ensuring transparency and taking advantage of the most innovative technologies. An interdisciplinary approach is needed to pursue this aim. The ongoing debate on science requires a collective effort from multiple disciplines and multiple areas of knowledge. Although the development of technology and technical skills is crucial, on the other hand, it is also important to pay attention to the legal aspects involved in the emergence of these new technologies, which – as such – pose novel legal challenges. It should not be forgotten that the openness of science is grounded in the “right to science”, enshrined in Article 27 of the Declaration of Human Rights. This presentation aims to investigate the future of Open Science, starting with the fundamental rights on which it is rooted.

        About the speaker:
        Ludovica Paseri is a PhD candidate at LAST-JD Joint International Doctoral degree program in “Law, Science and Technology”, at the Alma Mater Studiorum, University of Bologna, in cotutelle with the University of Luxembourg, faculty of Computer Science. She graduated in Law at the University of Turin and attended a postgraduate Master in Data Protection Law, at the Department of Law, University of Turin. Ludovica has completed the internship as trainee lawyer in an Administrative Law firm and her main research interests are Cloud Computing, EU Open Science policies, scientific data governance, privacy and data protection law.

        Speaker: Ms Ludovica Paseri (University of Bologna)
    • 12:45 PM
      Break
    • EOSC - Demonstration: EOSC Test Suite - Cloud Benchmarking and Validation go.egi.eu/egi2021-4 (Zoom Room 4)

      go.egi.eu/egi2021-4

      Zoom Room 4

      • 26
        Demo: EOSC Test Suite - Cloud Benchmarking and Validation

        In recent years, multiple EC funded projects are contributing to the adoption of cloud services in the European research sector. Helix Nebula Science Cloud (HNSciCloud) pioneered the development of a hybrid cloud model linking commercial cloud service providers and research organisations on-premises resources through the GÉANT network. Currently, projects like the Open Clouds for Research Environments (OCRE) leverage HNSciCloud experience in the exploitation of commercial cloud. Another project, CloudBank EU, proposes a model to monitor usage, scaling impact and broadening access of cloud in science, supporting multiple cloud contracts via a comprehensive set of user-facing and business operations functions. In more specific applications, initiatives such as ARCHIVER (Archiving and Preservation for Research Environments) promote FAIR through multiple technologies such as extreme data-scaling, network connectivity, service interoperability and business models, in a hybrid cloud environment to deliver end-to-end archival and preservation services covering the full research data management lifecycle.
        The aforementioned projects must integrate validation of services, in aspects such as performance, data sovereignty and cost in order for the cloud adoption to consolidate and become sustainable.
        In this context, CERN developed a tool that expanded the necessary testing and validation activities to help benchmarking services across vendors, providing working examples for researchers of their workloads, in multiple scientific domains. The approach consists of a modular and autonomous suite, based on open source technologies such as Terraform, Ansible, Kubernetes, Docker and CEPHS3 for results storing, that runs a set of benchmarks across multiple heterogeneous cloud stacks. The current test catalog offers validation tests on domains such as compute, storage, HPC, Machine Learning and network connectivity.
        The development of the suite started in early 2019, being improved continuously. The proposed demonstration will showcase the complete workflow, including configuration steps and results gathering. It will also discuss new developments and features foressen in the roadmap including flexible automated approaches with broader configuration options, increasing efficiency and resilience.

        Speaker bio's:
        Ignacio Peluaga Lozada: https://www.linkedin.com/in/ignacio-peluaga-lozada/?originalSubdomain=ch

        Shreyasvi Natraj: https://www.linkedin.com/in/nshreyasvi/?originalSubdomain=ch

        Speakers: Mr Ignacio Peluaga Lozada (CERN), Mr Shreyasvi Natraj (CERN)
    • 12:55 PM
      Chair Yoga go.egi.eu/egi2021-2 (Zoom Room 2)

      go.egi.eu/egi2021-2

      Zoom Room 2

    • 1:00 PM
      Lunch
    • 1:15 PM
      Lunch
    • EOSC - Demonstration: EOSC Portal - Onboard your resources and contribute to creating Europe’s trusted digital environment for research go.egi.eu/egi2021-4 (Zoom Room 4)

      go.egi.eu/egi2021-4

      Zoom Room 4

      • 27
        Demo: EOSC Portal - Onboard your resources and contribute to creating Europe’s trusted digital environment for research

        EOSC will become Europe’s virtual environment for all researchers to store, manage, analyse and re-use data for research, innovation and educational purposes. EOSC is intended to set off the ground by federating existing scientific data infrastructures and digital infrastructures for data exploitation that are now spread across disciplines and EU member states. This will make access to scientific data and other scientific outputs easier and more efficient.

        The EOSC Portal is part of the European Open Science Cloud implementation roadmap as one of the expected “federating core” services contributing to the implementation of the “Access and interface” action line. It provides a European delivery channel connecting the demand and supply sides of EOSC and its different stakeholders.

        The EOSC Portal is a gateway to information and resources in EOSC, providing updates on its governance and players, the projects contributing to its realisation, funding opportunities for EOSC stakeholders, relevant European and national policies, documents, and recent developments. The Portal welcomes the participation of providers that contribute to develop EOSC into a rich environment offering a wide range of services and resources for researchers through the EOSC Portal Catalogue & Marketplace.

        Onboarding and maintaining quality resources for research on the EOSC Portal is a great chance for providers to reach out to researchers across Europe and beyond. Providers receive support in the onboarding process, visibility on an evolving platform, updated statistics about usage and user feedback.

        This demonstration given by the EOSC Enhance team will present the benefits that providers get by uploading their services and resources into the EOSC Portal and will offer a tutorial to introduce its various features, such as the steps of the onboarding process, how to navigate the elements of the provider dashboard, how to add new resources or update existing resources and how to use the EOSC Portal API.

        Dr. George Papastefanatos is a senior researcher at Athena Research Center and affiliated with University of Athens- Department of Informatics & Telecommunications. He holds a Diploma
        in Electrical and Computer Engineering (2000) and PhD in Computer Science from the National Technical
        University of Athens (2009). George has more than 15 years of active involvement as a researcher and technical
        and project manager in several RTD projects, including projects related to Big Data engineering, big data
        visualization and analytics and the European Open Science Cloud activities. George was a member of the
        Architecture Working group of the European Open Science Cloud and he participates as a technical manager
        and researcher in a series of research projects which are developing EOSC (EOSC Future and EOSC-ENHANCE which develop
        and improve the functionality of the EOSC Portal for providers, eInfraCentral project, which offered the service catalogue
        functionality in the newly launched EOSC portal, CatRIS, which built the catalogue of services of Research
        Infrastructures and Neanias, whose goal is to develop and deliver sustainable services from the Atmosphere,
        Underwater & Space Communities). Also, George's expertise is in the area of big data management, data
        services on the cloud and data analytics and specifically in issues related to the web data and linked data
        modelling, data integration, archiving and preservation, information visualization and exploration. He has coauthored
        more than 60 papers in these areas, 1 book and 3 chapters in books, and 3 of his articles have been
        awarded best paper prizes in conferences.

        Speaker: George Papastefanatos (University of Athens)
    • Demonstration: Automating infrastructure and service audits with Governify go.egi.eu/egi2021-4 (Zoom Room 4)

      go.egi.eu/egi2021-4

      Zoom Room 4

      • 28
        Demo: Automating infrastructure and service audits with Governify

        In this demonstration we will present Governify (governify.io), a service agreement management framework that boosts the service governance by supporting audits in an automated way. It is composed of a set of integrated components that can be combined to create configurable architectures that adapt to each scenario. The governance platforms built with Governify gather evidences from multiple external sources in the organization (by means of their APIs) and provide visual dashboards to understand the current risks of not meeting the goals. The underlying agreeement model (iAgree) provides a uniform modelling approach in a wide range of domains: from Service Level Agreements in RESTful services, to Service Objective/Penalties and Rewards in IT Service Support Desks driven by humans, or Best Team Practices in Agile Development Teams. These holistic capabilities allow the definition of integrated metrics, goals and dashboards to create a common governance platform to drive the strategy of the organization. Governify provides a native micro-service architecture of RESTful Components that can be easily deployed and operated as containers in the infrastructure at choice.

        Speaker bio:
        Dr. Pablo Fernandez is Associate Professor at the University of Sevilla, Spain. He began his career as researcher in 2003 in the field of “automated contracting for web services” and has focused his research in the governance of services level agreements in organizations. He is member of the applied service engineering research group (ISA) publishing 50+ publications in conferences and journals and participating in different technology transfer and innovation projects to apply his research results to industrial settings in both the private and public sector. Since 2017 is a representative of the ISA Group at the governance board of the OpenAPI Initiative and has impulsed the creation of the SLA4OAI interest group to foster an extension to describe Service Level Agreement in RESTFul APIs integrated with the main OpenAPI Specification.

        Speaker: Pablo Fernandez (University of Sevilla)
    • EOSC - Presentations go.egi.eu/egi2021-4 (Zoom Room 4)

      go.egi.eu/egi2021-4

      Zoom Room 4

      Session Chair: Enol Fernandez (EGI)

      EOSC-Performance: find the most suitable EOSC site for your task
      Federating and orchestrating resources across the EOSC infrastructures
      The ESCAPE Data Lake as the bridgehead for the EOSC
      User experience with EOSC Compute Platform after 9 months

      Convener: Enol Fernandez (EGI.eu)
      • 29
        EOSC-Performance: find most suitable EOSC site for your task

        EOSC-Performance is a search-and-compare platform partially developed within the EOSC-Synergy project.

        Users and service providers can upload, search, and visually compare results from benchmarks for very diverse computing resources including cloud and HPC. This allows universities and computing centers to increase their visibility and scientists to identify the ideal provider for specific tasks.

        The EOSC-Performance service leverages OIDC and EGI-Check-In for authentication, Dynamic DNS service from EGI Federated cloud, provides API based on the OpenAPI 3 specification, and implies the SQAaaS approach from the EOSC- Synergy project. It is available in the EOSC Marketplace and it is going to be provided via the Infrastructure manager of Universitat Politècnica de València (UPV) for custom deployment.

        The service and typical use cases will be be presented in this contribution.

        Speaker bio:
        Borja Esteban Sanchis: https://submit.isc-hpc.com/contributors/129
        Valentin Kozlov: https://de.linkedin.com/in/valentin-kozlov-11b045b

        Speakers: Mr Borja Esteban Sanchis (Karlsruhe Institute of Technology), Valentin Kozlov (Karlsruhe Institute of Technology)
      • 30
        Federating and orchestrating resources across the EOSC infrastructures

        The INDIGO PaaS is a middleware that allows to federate distributed computing and storage resources using a TOSCA-based orchestration system.
        The development started during the H2020 INDIGO-DataCloud project (2015-2017) and then continued in the following projects: DEEP-HybridDatacloud, eXtreme-Datacloud and EOSC-Hub.
        During these years, the stability, reliability and scalability of the core services have been greatly improved. Moreover, several new features have been added: for example, there’s the possibility of deploying virtual computing environments (based on VM or container) with GPUs; the management of failures and timeouts has been improved; user management has been revised, introducing greater control over roles (regular user / admin). Another important update concerns the support for multi-tenancy and the integration with multiple authentication systems based on OpenID-Connect (in particular INDIGO-IAM and EGI-Checkin) to facilitate the adoption in different contexts, from INFN Cloud to EGI Federated Cloud. Moreover, it is now possible to federate not only private cloud infrastructures, such as Openstack, and public clouds such as Amazon AWS, and Mesos clusters, but also Kubernetes clusters: a plugin has in fact been added to manage Helm chart deployments. Finally, a web dashboard has been developed that greatly simplifies user interaction with PaaS services.
        The INDIGO PaaS Orchestrator is one of the services available in the EOSC marketplace (https://marketplace.eosc-portal.eu/services/infn.paas_orchestrator) and maintained by INFN.
        Currently it is being exploited in further projects, in particular C-Scale and EGI-ACE. The latter is implementing the compute platform of the European Open Science Cloud; our ambitious goals are now 1) to enlarge as much as possible the number of sites federated by the INDIGO PaaS in order to provide users with a transparent access to the distributed compute resources; 2) to provide high-level services ready to be deployed on demand by the end-users (e.g. JupyterHub as a service, Spark, TensorFlow, etc.). While in C-Scale the PaaS Orchestrator will be leveraged for supporting the EO use cases.

        Speaker bio:
        From 2003 to 2012 Marica Antonacci worked as a Software Engineer for an Italian private company developing software and technologically advanced solutions in the field of Earth Observation. In 2013 she started to work for INFN (Istituto Nazionale di Fisica Nucleare) gaining experience and skills in cloud computing technologies. She has been contributing to several European projects, like INDIGO-DataCloud, DEEP Hybrid-DataCloud, eXtreme-DataCloud, EOSC-Hub, EOSC-Pillar, EGI-ACE, etc. She has the role of system administrator of the production cloud site at the ReCaS datacenter (INFN-CLOUD-BARI) and is responsible for its integration in the EGI Federated Cloud.

        Speaker: Marica Antonacci (INFN)
      • 31
        The ESCAPE Data Lake as the bridgehead for the EOSC

        Experiments and scientists, whether in the process of designing and building up a data management system or managing multi-petabyte data historically, gather in the European Science Cluster of Astronomy & Particle physics ESFRI research infrastructures (ESCAPE) project to address computing challenges by developing common solutions in the context of the EOSC.
        A modular ecosystem of services and tools constitutes the ESCAPE Data Lake, which is exploited by flagship ESFRIs in Astro-particle Physics, Electromagnetic and Gravitational-Wave Astronomy, Particle Physics, and Nuclear Physics to pursue together the FAIR and open-access data principles.
        This infrastructure fulfils the needs of the ESCAPE community in terms of data organisation, management, and access, and dedicated assessment exercises demonstrated its robustness.
        As a result, collaborating sciences are choosing their reference implementations of the various technologies among the proposed solutions.
        A variety of challenges and specific use cases boost ESCAPE to carefully take into account both user and infrastructure perspectives, and contributed to successfully conclude the pilot phase beyond expectations, embarking on a like-production prototype stage.
        The ongoing phase of the project aims at consolidating the functionalities of the services, e.g. integrating token-based AuthN/Z or deploying a tailored content delivery and caching layer, and at simplifying the user experience. Specifically for this reason, a considerable effort is being devoted towards a DataLake-as-a-Service whose goal is to provide the end-user with a Notebook ready-to-be-used and fully integrated with the Data Lake.
        ESCAPE milestones achieved during the length of the project represent a fundamental accomplishment under both sociological and computing model aspects for different scientific communities that should address upcoming data management and computing challenges in the next decade.

        Speaker bio:
        Riccardo Di Maria is leading the effort on novel data access technologies for future distributed storage infrastructures, referred to as the Data Lake, being prototyped in the ESCAPE European-funded project and leveraging Worldwide LHC Computing Grid (WLCG) technologies. ESCAPE aims to integrate research facilities of Astro-particle and Particle Physics, Electromagnetic and Gravitational-Wave Astronomy, and Nuclear Physics into a common data infrastructure in the context of the European Open Science Cloud (EOSC), pursuing the FAIR and open-access data principles.

        Speaker: Riccardo DI MARIA (CERN)
      • 32
        User experiences with the EOSC Compute Platform after 9 months

        The new EGI’s flagship project started in January 2021 proposes to deliver the European Open Science Cloud Compute Platform and expand the supply-side, contribute to the implementation of the EU Data Strategy and the EOSC Data Commons to support the Green Deal, Health, Fundamental Research and Social Sciences and Humanities data spaces, integrate the EOSC Compute Platform in the EOSC Portal and the EOSC Core, contribute to the realization of a Global Open Science Cloud, and expand the demand-side and facilitate cross-domain sharing and exploitation of research datasets.

        In this presentation we will focus more on how the project is contributing to facilitate cross-domain sharing and the exploitation of research datasets.

        The cross-domain use cases supported by EGI-ACE included:

        • 13 Data Spaces addressing the following disciplines: Health and Medicine (4); Climate Research (2); Energy and Physical Sciences (2); Environmental Sciences (4); and Social Sciences and Humanities (1).
        • 7 international Early Adopters scientific communities aiming at expanding the initial capabilities offered by the project becoming new Data Space providers.

        To support the 13 Thematic Data Spaces and the 7 Early Adopters cross-domain use cases the project created distributed Competence Centres across the NGIs, user communities, service and technology providers of the EGI Collaboration. Through these distributed Competence Centres the project provided the technical and the training support for supporting the integration of the scientific use cases in the EOSC Compute Platform.

        To complement the engagement process and identify new emerging scientific areas and underrepresented disciplines, the project opened a call for use cases. By the time of writing of this abstract, during the first two cut-off dates of the EGI-ACE open call, more than 10 scientific use cases were selected and supported by the project. For each of the supported scientific use cases the project offered access to infrastructure and platform services, dedicated user support and training. The services and the resources offered to the use cases are sponsored by the European Commission and various national funding agencies and are free to access to the selected use cases.

        In this presentation the status of the integration plans of the Thematic Data Space providers, Early Adopters and applications in the EOSC Compute Platform will be presented along with the lessons learnt after 9 months of project.

        Speaker bio:
        Giuseppe works as Community Team Lead at the EGI Foundation. One of his main activities is to establish strategies for the retention, development and growth of the network research projects and initiatives collaborating with EGI, of the community of users of EGI services, and for managing the EGI Engagement and Support Annual Plan and the Training Annual Plan.
        Since 2004, both at National and European level, he has worked as technologist for the Italian National Institute of Nuclear Physics (INFN) division of Catania in distributed computing projects co-funded by the European Commission.

        During these years, he has matured strong skills and competences on Grids and Clouds technologies, working on ICT scientific developments for supporting both emerging and already established VRCs. Giuseppe holds a MSc in Computer Science Engineering from the University of Catania (Italy).

        Speaker: Giuseppe La Rocca (EGI.eu)
      • 33
        EOSC DIH: Bridging Industry with the EOSC

        The EOSC DIH is an international and multi-partner cooperation that supports private companies in easily accessing the digital technologies, services and human capital offered by the EOSC. It combines four pillars to help companies become more competitive:
        Piloting and co-design: where products or services are designed and/or tested.
        Technical access: High-throughput, High-performance and cloud computing; machine learning; research data; data management services, as well as several tools and applications available for re-use.
        Training and support: technical consultancy, business coaching and funding search.
        Visibility: inclusion of company solutions in the EOSC Marketplace, print and online material and exploitation of the EOSC networking and ecosystem opportunities such as joining online communities, event attendance and potential future projects.

        There are already several EOSC DIH pilots developing innovative solutions and bringing them to market in different domains such as sport and biomedicine, cybersecurity, manufacturing, transport, agriculture or environment. More information about current pilot and success stories can be found here.

        Speaker bio: https://www.linkedin.com/in/marcinplociennik/?originalSubdomain=uk

        Speaker: Marcin Plociennik (ICBP)
    • Envisioning the Future - Workshop: Joining forces for a robust digital landscape: European Cloud open source and standardization effort go.egi.eu/egi2021-1 (Zoom Room 1)

      go.egi.eu/egi2021-1

      Zoom Room 1

      Cloud technologies are key enablers to unleash the potential of the European data economy and ensure its industrial competitiveness. The current cloud ecosystem needs to be reinforced and enlarged to embrace initiatives and efforts beyond the core European Cloud Computing community for aligning on a common strategy, while coordinating on its implementation.

      Roadmapping, outreach and community building activities lay an essential foundation for exploring and strengthening the European Cloud Computing ecosystem, however unifying support activities that facilitate the uptake and deployment of next generation cloud computing technologies also prove to be crucial to advance in the shorter-term the digital transformation of Europe. Open-source and (pre-)standardisation activities are pivotal enablers of technology and solution adoption and even ecosystem development. There are nevertheless multiple challenges to overcome on the path to exploitation, uptake and deployment of cloud computing technologies.

      In this session, the panel of experts from public and private organisations will discuss these potential roadblocks and cite their recommendations especially for SMEs and entrepreneurs, to help them identify acceleration paths to support their exploitation plans, speeding up their cloud offering to reach the market. The session will delve deep into discussions on the legal, contractual, financial, and business enablers that must be in place and known before
      being able to conceive and plan a sustainable value proposition.

      Speaker suggestion:

      Representative from Eclipse Foundation
      Representative from “open source” Foundation : eg, CNCF, OpenStack,
      Representative from a Start-up accelerator
      Representative from a tech Start-up
      Legal professional specializing in tech contracts or representative from the policy makers

      Convener: Jurry Mar, de la (T-Systems International GmbH)
      • 34
        Joining forces for a robust digital landscape: European Cloud open source and standardization effort

        Cloud technologies are key enablers to unleash the potential of the European data economy and ensure its industrial competitiveness. The current cloud ecosystem needs to be reinforced and enlarged to embrace initiatives and efforts beyond the core European Cloud Computing community for aligning on a common strategy, while coordinating on its implementation.

        Roadmapping, outreach and community building activities lay an essential foundation for exploring and strengthening the European Cloud Computing ecosystem, however unifying support activities that facilitate the uptake and deployment of next generation cloud computing technologies also prove to be crucial to advance in the shorter-term the digital transformation of Europe. Open-source and (pre-)standardisation activities are pivotal enablers of technology and solution adoption and even ecosystem development. There are nevertheless multiple challenges to overcome on the path to exploitation, uptake and deployment of cloud computing technologies.

        In this session, the panel of experts from public and private organisations will discuss these potential roadblocks and cite their recommendations especially for SMEs and entrepreneurs, to help them identify acceleration paths to support their exploitation plans, speeding up their cloud offering to reach the market. The session will delve deep into discussions on the legal, contractual, financial, and business enablers that must be in place and known before
        being able to conceive and plan a sustainable value proposition.

        Draft agenda:
        14:00-14:10 - Introduction: Dr. Giovanni Rimassa, Coordinator HUB4CLOUD, CIO Martel Innovate
        14:15- 15:05 - Expert Panel on the session topic
        15:10- 15:20 - Guest talk - European OpenSource efforts and activities (tbd)
        15:20- 15:30 - Q and A and closing

        Name of our panelists:
        Luca Bolognini, ICT Legal Consulting
        Brian King, Cloud Development Tools Community Manager, ECLIPSE FOUNDATION
        Alberto P. Marti, VP, Open Source Community Relations, OpenNebula
        Dr. Arne Berre, Chief Scientist, Software and Service Innovation Division, SINTEF

        The panel will be moderated by Dr. Giovanni Rimassa.

        Speakers: Eclipse Foundation, CNCF, Openstack Foundation, Startup accelerator, Dr Giovanni Rimassa (Martel Innovate GmbH), Legal expert in technology contracts, Amrita Prasad (Martel Innovate)
    • Innovating Services Together - Presentations go.egi.eu/egi2021-2 (Zoom Room 2)

      go.egi.eu/egi2021-2

      Zoom Room 2

      Convener: Smitesh Jain (EGI.eu)
      • 35
        Innovation Management in EGI Federation

        Everybody talks about Innovation.. but what does it really mean? Everybody has a definition for innovation and quite often there is no commonality between these varied definitions. So it becomes even more difficult to put into practice Innovation Management which is the systematic promotion of innovations in organizations through planning, organization, management and control. To make it more difficult it becomes even more complicated with the scope of activities under Innovation Management which ranges from trend scouting to idea generation and all the way to IPR management.
        The main objective of this session is to try to remove the fuzziness and messiness that surrounds Innovation Management and its implementation in a federated environment. The session will begin with delineating the differences between Innovation Management in collaborative projects and Innovation Management in an organization. The presentation will then discuss the practicalities of how the EGI Foundation is implementing Innovation Management. It will also discuss how the concept of Open Innovation will be exploited to involve the EGI community in the Innovation Management process. The presentation will also share learnings and best practices in implementing an Innovation Management process.
        The session aims to increase the collaboration between the wider EGI community by discussing the approach of open innovation. By sharing best practices the session also provides inspiration and key learnings for participants to implement or improve their Innovation Management process.

        Speaker bio:
        Smitesh joined EGI Foundation in April 2021 as Innovation Management Specialist in the Strategy and Innovation Team. He has a MSc. in Management of Technology from the Technische Universiteit Delft, Netherlands. Smitesh is working on developing and evolving Innovation Management processes and systems at EGI alongside managing innovation in ongoing projects. Smitesh is a certified Design Thinking trainer and Lego Serious Play facilitator.

        Speaker: Smitesh Jain (EGI.eu)
      • 36
        On demand data analysis tools for the EOSC Resource Catalogue

        OpenAIRE (www.openaire.eu) aims to establish an open and sustainable scholarly communication infrastructure responsible for the overall management, analysis, manipulation, provision, monitoring and cross-linking of all research outcomes. One of the contributions of OpenAIRE to the European Open Science Cloud (EOSC) is its research graph (https://graph.openaire.eu), one of the largest open scholarly record collections worldwide, which constitutes the EOSC Resource Catalogue. Conceived as a public and transparent good, populated out of data sources trusted by scientists, the EOSC Resource Catalogue aims at bringing discovery, monitoring, and assessment of science back in the hands of the scientific community.
        It includes metadata records and links (i) collected from 70k+ scholarly communication sources from all over the world, including Open Access institutional repositories, data archives, journals; (ii) inferred by data mining algorithms, and (iii) provided by users of the OpenAIRE portals thanks to the Link functionality.
        OpenAIRE is already using the catalogue to power its portals and to support researchers, funders, organisations, research communities and infrastructures at discovering and tracking research products. The catalogue is also openly accessible via data dumps and APIs (https://develop.openaire.eu), so that its content can be used by any researchers for their own research activities.
        In order to ease the usage and analysis of the EOSC Resource Catalogue, this presentation proposes an integration with the EGI notebooks on demand (https://marketplace.egi.eu/44-notebooks).
        EGI Notebooks is a browser-based tool for interactive analysis of data using EGI storage and compute services based on the JupyterHub technology. Notebooks are offered on-demand to single researchers or research communities.
        The idea is to offer EGI notebooks capable of requesting “slices” of the EOSC Resource Catalogue that are relevant for the end user and support the definition of functions for the analysis of such slices.
        Thanks to the EGI Notebook, the end users will be able to analyse the EOSC Resource Catalogue on top of the stable and scalable infrastructure of EGI.
        For example, if the end users want to analyse the outputs of H2020 projects, the notebook will download the dump of H2020 research outputs from Zenodo (https://doi.org/10.5281/zenodo.4559725) and execute a number of predefined functions to give an overview of the data to the end-users.
        The EGI notebook should also be able to address requests for slices of the catalogue that are not already published on Zenodo by OpenAIRE. In such cases, the users should be able to define the selection criteria that will be applied to the full dump (https://doi.org/10.5281/zenodo.3516917) or used to build a query to the OpenAIRE or EOSC discovery API.

        Speaker:
        Michele De Bonis is born in Italy. Computer Engineer and PhD student at ISTI-CNR aiming to study Deep Learning techniques for graph processing. Currently responsible for the implementation and the maintenance of the OpenAIRE services.

        Speaker: Michele De Bonis (CNR-ISTI)
      • 37
        mytoken - secure and long term authentication (without refresh tokens)

        Background

        OpenID Connect (OIDC) is the technological basis of many modern Authentication and Authorisation Infrastructures, which are currently being used and established in multiple European projects. Also, the non-academic sector (e.g. Microsoft, Google, Apple, IBM) moved to OIDC.

        Despite OIDC being mostly used within web browser based applications, support for the command line and for API usage are required for complex workflows. The oidc-agent
        tool was our first step to enabling OIDC's "Access Tokens" on the command line. Access Tokens are short lived, with a life-time of usually one hour.

        Use-cases that involve long running jobs, however, require authentication capabilities throughout their entire runtime, thereby challenging the security concept of short lived tokens.

        Mytoken

        This contribution addresses complex scenarios (e.g. compute jobs), in which access tokens need to be obtained over extended time spans, e.g. to load and store data or to access other resources.

        We introduce a client-server solution and a new token type, called mytoken. Mytokens are easy to use, (can be) long lived, and allow limiting the functionality of the token, to address security concerns that arise from long living tokens.

        The mytokens themselves are very flexible, as they can be
        - used for easily obtaining access tokens on any device
        - easily transferred to other devices
        - created non-interactively from another mytoken
        - easily passed around without giving up security
        - restricted according to the use case.

        Mytokens may have capabilities and restrictions. Capabilities define
        well-defined actions for which a mytoken may be used (e.g. obtain an
        Access Tokens, obtain another mytoken), while restrictions may be used for
        a fine grained access control, for example:

        • Time range in which a mytoken may be used
        • Location (IP, Geo-IP) from where the mytoken may be used
        • OIDC (AT scope, audience) to specify what kind of Access Tokens may be
          obtained
        • Number of usages for a specific action

        Mytokens can contain lists of restrictions, which allows to easily
        define a mytoken, that could:
        1. Allow job submission in the first hour after creation of the mytoken
        2. Allow data access in the first two hours after creation
        3. Allow nothing for one day
        4. Allow data access (store output) between 24 and 36 hours after creation.

        Speaker: Marcus Hardt (KIT-G)
      • 38
        DIGITbrain – a new approach to deliver Digital Twin technologies

        DIGITbrain – a new approach to deliver Digital Twin technologies
        The goal of the DIGITbrain Project (funded by the EC, grant Nr. 952071) is to enable
        customised Industrial Products, means manufacturing machines and processes, and facilitate
        cost-effective distributed and localised production for manufacturing SMEs by applying
        modelling, simulation, optimisation, analytics, and machine learning tools and augmenting
        the concept of Digital Twins with a memorising capacity.
        Building on the CloudiFacturing Solution (and the Digital Agora, publicly known as emGORA
        Workspace), a platform for offering and consumption of digital services, the DIGITbrain
        Platform will extend the CloudiFacturing solution with new capabilities, like publishing,
        composing and orchestration of data, models, algorithms, and resources. The new solution
        integrates technologies like Clouds, FIWARE, co-simulation, Machine Learning, etc., in order
        to develop a MaaS (Manufacturing as a Service) business model to be implemented by the
        Digital Innovation Hubs (DIHs) later on.
        During the lifetime of the DIGITbrain project, two Open Calls are focused on attracting third
        Parties (especially manufacturing and ISV SMEs and mid-caps) to perform their experiments
        and to make SMEs more competitive by transferring innovative solutions into the wider
        manufacturing community. Experiments will be implemented in close collaboration with
        Digital Innovation Hubs. Currently the first Open Call is in evaluation; the next Open Call will
        be launched in July 2022.
        Technically DIGITbrain is based on the previous mentioned Digital Agora
        (workspace/marketplace), additionally further services are developed, like automated
        deployment, execution on Edge and Cloud, HPC resources and the memorising capability of
        the “Digital Product Brain” in order to monitor, explore and learn from the implemented
        Digital Twins. The “Digital Product Brain” itself is the coordinating building block of the
        DIGITbrain Solution and it aims to configure, manage, and monitor Digital Twins and their
        DMA (Data, Model, Algorithm) Tuples (representation of individual behaviours of Industrial
        Products).
        Practically as a user, you just have to author, publish, and compose your assets (Data,
        Model, Algorithm) for using them or offering them as a service on the DIGITbrain solution.
        Main advantages of the DIGITbrain concept are
        • The clear separation of Data, Model and Algorithms (Assets) in the DIGITbrain
        Concept offers reusable building blocks to compose Digital Twins more rapidly.
        • For transparency, analysis and further usage, those assets are described and
        published by metadata as well
        • Provision of authoring tools and algorithms should lower the barriers and ease the
        access to the DIGITbrain solution for the involved SME’s and DIHs.
        Together with the DIHs, innovative services will be developed, published on the Digital Agora
        and delivered to European SMEs. Additional consulting services, provided by the involved
        DIHs, the contributing SMEs and core partners in the project, supports the adaption of the new technical services and are core of the envisaged community building process. Together
        with the DIHs and based on the DIGITbrain solution, the new business model
        “Manufacturing as a Service” MaaS will be explored, investigated, and established later.

        Speaker bio: https://www.westminster.ac.uk/about-us/our-people/directory/kiss-tamas

        Speaker: Tamas Kiss (University of Westminster, London, UK)
      • 39
        ScienceMesh + JupyterLab: Collaborative Data Science services in scientific use cases and in business across different fields of study

        Data Science became famous in main stream about a decade ago: after Harvard Business Review article which coined Data Scientist as “The Sexiest Job of the 21st Century”. In business it is defined as “using data to increase competitive advantage”.
        ScienceMesh, developed in CS3MESH4EOSC project, creates the Federated Scientific Mesh providing federated sharing of data across different sync-and-share services, federated use of applications (such as collaborative document editing, data archiving, and data publishing), fast transfer of large datasets and remote data analysis (distributed Data Science environments).
        As all scientific disciplines nowadays are based on data analysis and Distributed Data Science Environments can support research in all fields of study. It will also support Data Science in the business context in all sectors. In recent report Critical Capabilities for Data Science and Machine Learning Platforms (March 2021) Gartner predicts, that in near future collective intelligence in Data Science and cloud-based AI infrastructure will be among key factors for competitive advantage.
        For ScienceMesh Distributed Data Science Environments, along with Software Mind (part of Ailleron group, a global IT service provider based in Poland, delivering skilful managed teams for even most demanding projects), it was developed the JupyterLab extension, integrating with ScienceMesh – file browsing and additional share and collaboration functionalities for notebooks and resources across federated cloud are now possible in JupyterLab environment. Collaborative Data Science is being now used in products from Finance, IoT, Earth Observation, SmartCities and Pharma, but it is present in virtually every business.
        Jupyter Notebook has become No1 platform used by data scientists to build interactive applications and to work with big data and AI. It is a free, open-source, interactive web-based tool which researchers can use to combine software code, computational output, explanatory text and multimedia resources in a single document. Jupyter has exploded in popularity over the past couple of years, with an enthusiastic community of user–developers.
        In this talk, the relevance and benefits of ScienceMesh Distributed Data Science Environments will be presented, starting from two scientific use cases (High Energy Physics and Earth Observation), along with various business-related scenarios.
        In CS3MESH4EOSC project, SoftwareMind is supporting science innovation by providing the expertise on microservices architecture, integration, DevOps, agile software development process and Data Science. It leads tasks on Reference interoperability platform and distributed Data Science environments. In this talk, the speaker will show also how this is as a part of the strategy of growing the usage of application services in the cloud, microservice-based architectures, Data Science, Big Data integration and analytics.

        Speaker bio:
        Experienced System Architect and R&D Project Manager with 20+ years of enterprise software design and development. Founder and leader of Big Data Lab, focused on R&D, Data Science, data driven innovation and high quality agile software development. He is now involved in CS3MESH4EOSC project (https://cs3mesh4eosc.eu/), leading tasks on Reference cloud interoperability platform and distributed Data Science environments (the project creates Science Mesh integrated with EOSC: https://cs3mesh4eosc.eu/index.php/science-mesh). He developed Big Data solutions before it became mainstream. In the years 2005-2008 he was involved in development of technology for the first web-scale Semantic Web startup: garlik.com - from R&D in alpha phase to commercial launch browsing 4billion web pages. Software Mind’s team he was a part of - and finally he lead - started using Hadoop in February 2006, as one of the first companies in the world. He participated in and lead many commercial projects which included Big Data, high volume and high velocity solutions, in various sectors: telco operators, international telco interoperability hubs, banks and financial institutions or Content Delivery Network providers. He was a Work Package leader and provided Big Data architecture in in a number of EU-funded research projects. He was Chief Software Engineer and integration WP leader in ROBUST (2010-2013). He was a WP leader on technological infrastructure in EU IP WeSenseIt (2012-2016), responsible for designing and implementing a scalable architecture of sensor/IoT platform, management of huge scale geospatial data, scalable backend for mobile apps and cloud-based infrastructure. In Seta (2016-2019) he leads a Work Package on creating Big Data infrastructure for organizing, monitoring and planning multimodal mobility in large metropolitan areas. His current focus in Big Data innovation is state of the art solution for management of geo-located data and low-latency services, including GPU-based acceleration of geospatial indexes. He has experience in all phases of project lifecycle including requirements gathering and analysis, architectural analysis and design, data modelling, implementation, deployment, coordinating and mentoring.

        Speaker: Marcin Sieprawski (Software Mind)
      • 3:25 PM
        Q&A
    • 3:30 PM
      Break Pïmp your Badge!

      Pïmp your Badge!

      https://indico.egi.eu/event/5464/page/94-pimp-your-badge
    • Delivering Services and Solutions - Workshop: Orchestrating Data-intensive Applications on Federated Hybrid Infrastructures with the SODALITE Framework go.egi.eu/egi2021-2 (Zoom Room 2)

      go.egi.eu/egi2021-2

      Zoom Room 2

      Data-intensive complex applications, which include microservices, computationally intensive batch jobs, and sophisticated interaction with the external environment, demand for heterogeneous computational infrastructures that range from (public, private, and federated) Cloud to HPC and Edge. In this context, a crucial problem is to facilitate the work of DevOps teams in 1) the conception of the right operational architecture for the application, 2) its transformation into infrastructural code that automates its deployment, taking into account the peculiarities of each of the diverse infrastructures involved in this, and 2) its operation. The SODALITE framework aims at addressing this scenario. In pursuit of that objective, the SODALITE meta-model is developed on top of the TOSCA standard, which enables the interoperable description of application and infrastructure components.

      The goal of this tutorial is to provide a beginner level hands-on training of the SODALITE framework. The tutorial is in three parts. The first part will briefly introduce the Infrastructure as Code (IaC) approach and the IaC languages that the SODALITE uses. IaC is the process of managing and provisioning IT infrastructures such as storages and networks through machine-readable definition files. It is rapidly emerging the de facto approach for provisioning and managing IT infrastructure. The second part of the tutorial will provide an overview of the SODALITE framework, focusing on its smart modeling environment (e.g., IDE and context-aware content-assistance services) and orchestration capabilities (e.g., deployment, monitoring, data management, and authentication/authorization). The final phase of this tutorial, using two case studies, will demonstrate the design of a defect and error free deployment model with ease using the SODALITE smart IDE, and the enactment of the designed model on Cloud (EGI and Openstack) and HPC clusters.

      Convener: Indika Priyantha Kumara Dewage (Tilburg University)
      • 40
        Orchestrating Data-intensive Applications on Federated Hybrid Infrastructures with the SODALITE Framework

        Data-intensive complex applications, which include microservices, computationally intensive batch jobs, and sophisticated interaction with the external environment, demand for heterogeneous computational infrastructures that range from (public, private, and federated) Cloud to HPC and Edge. In this context, a crucial problem is to facilitate the work of DevOps teams in 1) the conception of the right operational architecture for the application, 2) its transformation into infrastructural code that automates its deployment, taking into account the peculiarities of each of the diverse infrastructures involved in this, and 2) its operation. The SODALITE framework aims at addressing this scenario. In pursuit of that objective, the SODALITE meta-model is developed on top of the TOSCA standard, which enables the interoperable description of application and infrastructure components.
        The goal of this tutorial is to provide a beginner level hands-on training of the SODALITE framework. The first phase of the tutorial will provide an overview of the SODALITE framework,
        focusing on its smart modeling environment (e.g., IDE and context-aware content-assistance services) and orchestration capabilities (e.g., deployment, monitoring, data management, and authentication/authorization). The second phase is a hands-on tutorial, which teaches the participants how to design a defect and error free deployment model with ease using the
        SODALITE smart IDE and orchestrate the designed model on the SODALITE testbed. Participants interested in conducting the hands-on session tutorial (i.e. the IDE demonstration)
        should install beforehand the SODALITE IDE (compatible with Windows, MacOS and Linux) by following the installation procedure described at:
        https://github.com/SODALITE-EU/ide/blob/master/README.md

        Presenters:
        Jesús Gorroñogoitia, BS in Theoretical Physics from the Universidad Complutense de Madrid (UCM), with a Master in Condensed Matter and Statistics Physics by UNED. In Atos Research & Innovation (ARI), he is senior researcher and software architect in the Advance Parallel Computing Lab, specialized in HPC orchestration. Before, he was the ARI Research Line Expert on Software Engineering, working on Service Oriented Computing (SOC), Model Driven Development (MDD), Autonomous Computing, or Semantics.

        Dragan Radolović graduated from University of Ljubljana in 1998 with a B.Sc. thesis titled “Image database queries based on color information”. His background includes work in the
        financial field (banks, financial institutions) and other software industry as a software architect and project manager. His main points of interest are Cloud architectures. Currently he works at XLAB Research as a project manager and work package leader on the SODALITE H2020
        project.

        Zoe Vasileiou has attended the school of Computer and Computer Engineering of the University of Thessaly, Greece from which she graduated in 2013. She also received a Master's from the same department with the title "Science and Technology of Computer Engineering, Telecommunications and Networks". Next, she worked for 5 years as a Software Engineer in Scientific Games Corporation which provides gambling products across the globe. Now, she is a
        PhD student in the School of Informatics of the Aristotle University of Thessaloniki in the field of Explainable Artificial Intelligence. At the same time, she is working at CERTH research center in Greece where she mainly has engaged in research projects related with Semantic Web technologies and cloud computing. Her research interests include Explainable AI, Knowledge Representation, and Ontology Engineering.

        Indika Kumara is a post-doctoral researcher at the Jheronimus Academy of Data Science and Tilburg University, the Netherlands. He received his PhD degree from Swinburne University of Technology, Australia. His research interests include service-oriented computing, cloud computing, software performance engineering, self-adaptive systems, and Quantum computing.

        Speakers: Indika Priyantha Kumara Dewage (Tilburg University), Mr Dragan Radolović (XLAB Research), Mr Jesús Gorroñogoitia (ATOS), Mrs Zoe Vasileiou (Information Technologies Institute, Centre for Research and Technology Hellas)
        • a) Introduction to SODALITE
        • b) IaC and Services of SODALITE
        • c) Hands-on Session using SODALITE Tools to Design and Deploy a Heterogenous Application
    • Delivering Services and Solutions - Workshop: The EGI Federated Cloud: benefits for service providers and customers go.egi.eu/egi2021-3 (Zoom Room 3)

      go.egi.eu/egi2021-3

      Zoom Room 3

      This session provides an overview of the computing services of EGI that deliver a distributed federated cloud infrastructure to perform any kind of data analytics for research and innovation.

      The EGI service portfolio provides you various computing solutions to match your needs: Virtual Machine based computing for long-running services and for data analytics platforms; container orchestration powered by Kubernetes and Docker; facilities for massively parallel workloads. During this session we will describe the benefits the federation brings to both service providers and customers. The session will also feature research communities, including EGI-ACE thematic services, to report on their success stories in using these services in real-life research workflows.

      Convener: Enol Fernandez (EGI.eu)
      • 41
        The EGI Federated Cloud: benefit for service providers and customers

        This session provides an overview of the computing services of EGI that deliver a distributed federated cloud infrastructure to perform any kind of data analytics for research and innovation.

        The EGI service portfolio provides you various computing solutions to match your needs: Virtual Machine based computing for long-running services and for data analytics platforms; container orchestration powered by Kubernetes and Docker; facilities for massively parallel workloads. During this session we will describe the benefits the federation brings to both service providers and customers. The session will also feature research communities, including EGI-ACE thematic services, to report on their success stories in using these services in real-life research workflows.

        Speaker: Enol Fernandez (EGI.eu)
      • 42
        EGI Check-in for the Federated Cloud

        This session will provide an overview of the EGI Federation Registry tool that allows service providers to manage the connection of their OpenID Connect and SAML based services to Check-in, the authentication, authorisation and user management service for the EGI infrastructure. The EGI Federation Registry provides a web interface which covers the whole service lifecycle, including the initial registration, reconfiguration and deregistration of a service. Cloud resource owners can use the Federation Registry to connect their compute services to the EGI Check-in authentication and authorisation service in order to join the Federated Cloud as providers and enable research communities to access their resources.

        Speakers: Andreas Kozadinos (GRNET), Nicolas Liampotis (GRNET), Valeria Ardizzone (EGI.eu)
      • 43
        Deploying virtual elastic clusters on the EGI Cloud Compute

        In this presentation, we will introduce you EC3, a tool able to deploy virtual elastic clusters (i.e. Kubernetes or SLURM clusters, among others) on top of the EGI Cloud Compute. You will know more details about the tool, its main capabilities, its interfaces (both CLI and Web portal) and how to start using it. Don't miss the session!

        Speakers: Dr Amanda Calatrava (UPVLC), Miguel Caballer (UPVLC)
      • 44
        Experience with commercial clouds in Science

        T-Systems was able to collect various experience working with the science community through e.g. Helix Nebula, Indigo-Datacloud, Copernicus and currently the OCRE framework.
        The presentation will share some interesting findings and lessons learned, what is required to make good use of and in which scenarios science can benefit from commercial clouds.

        Speaker: Jurry Mar, de la (T-Systems International GmbH)
      • 45
        Towards an e-infrastructure for plant phenotyping

        Various initiatives have helped to structure the european phenotyping landscape (EMPHASIS, EPPN) and enable researchers to use facilities, resources and services for plant phenotyping across Europe. Among these services there is a need to build a federated and interoperable e-infrastructure allowing researchers to share and analyze phenotyping data. The European Network Infrastructure (EGI) can greatly contribute to the deployment of such an e-infrastructure, as demonstrated by the EAP program.

        Speaker: Vincent Negre (INRA)
      • 46
        Discussion
    • EOSC - Workshop: EOSC Future and EOSC-Core go.egi.eu/egi2021-4 (Zoom Room 4)

      go.egi.eu/egi2021-4

      Zoom Room 4

      During this session we'll discuss how the EOSC Future project facilitaties EOSC operation and its architecture.

      We'll also take a look at the current status and planned developments in EOSC Core, and we'll give an overview of the status and planned evolution of the EOSC onboarding process.

      Conveners: Diego Scardaci (EGI.eu), Gergely Sipos (EGI.eu), Matthew Viljoen (EGI.eu)
    • Innovating Services Together - Presentations: The EGI DIH, a new space for the engagement with Industry go.egi.eu/egi2021-1 (Zoom Room 1)

      go.egi.eu/egi2021-1

      Zoom Room 1

      The EGI Digital Innovation Hub (“EGI DIH”) is a virtual space where companies and technical service providers meet to test solutions before investing, offering different services on advanced computing to help companies in digitalization and improving productivity. The main focus of the DIH is the coordination and development of both human and technical support for business-oriented pilots and collaborations.

      The EGI DIH acts as a one-stop-shop to provide technical assets, knowledge, expertise and support on business, market, and finance to implement innovation. It aims at 1)Help and support companies in the adoption/ test/validation of advanced computing technologies to become more competitive, 2) Provide an open and innovative environment for experimentation and innovation, increasing the innovation among the EGI Federation members and facilitating their relationship with business sectors for a wider economic and societal impact.

      This session will introduce the EGI DIH as one of the business models of the EGI Business Model Innovation Program with the participation of several use cases to share their experiences working with computing services and the knowledge generated, as well as strategic partnerships with other innovative initiatives to collaborate.

      The session is oriented to the EGI community, Industry, DIH networks and other innovation agents with interest in innovation and collaborative networks for digital transformation.

      Convener: Elisa Cauhe (EGI.eu)
      • 51
        Session Intro and Overview
        Speaker: Elisa Cauhe (EGI.eu)
      • 52
        The EGI DIH, a new space for the engagement with Industry

        The EGI Digital Innovation Hub (“EGI DIH”) is a virtual space where companies and technical service providers meet to test solutions before investing, offering different services on advanced computing to help companies in digitalization and improving productivity. The main focus of the DIH is the coordination and development of both human and technical support for business-oriented pilots and collaborations.

        The EGI DIH acts as a one-stop-shop to provide technical assets, knowledge, expertise and support on business, market, and finance to implement innovation. It aims at 1)Help and support companies in the adoption/ test/validation of advanced computing technologies to become more competitive, 2) Provide an open and innovative environment for experimentation and innovation, increasing the innovation among the EGI Federation members and facilitating their relationship with business sectors for a wider economic and societal impact.

        This session will introduce the EGI DIH as one of the business models of the EGI Business Model Innovation Program with the participation of several use cases to share their experiences working with computing services and the knowledge generated, as well as strategic partnerships with other innovative initiatives to collaborate.

        The session is oriented to the EGI community, Industry, DIH networks and other innovation agents with interest in innovation and collaborative networks for digital transformation.

        Speaker: Elisa Cauhe (EGI.eu)
      • 53
        Selected EGI DIH Business Case: Mathworks
        Speaker: Shubo Chakrabarti (MathWorks)
      • 54
        Business Pilots via Projects: EUH4D Pilot - ioT SESOD Binary Oy
        Speaker: Andrei Costin
      • 55
        Business Pilots via Projects: EOSC DIH Pilot supported by EGI-ACE - Trango
        Speaker: Carlos Rubia
      • 56
        DIH Collaborations: emGORA Workspace – A community-based approach to deliver software services for manufacturing engineering

        emGORA workspace – a community-based approach to deliver software services for manufacturing engineering
        In the past 10 years, the community-based approach is enjoying increasing popularity. The European Single Market offers many opportunities for large companies that are well
        equipped with financial and human resources, but where small and individual players unfortunately fall by the wayside. But how can the small and medium-sized ones be supported?
        The emGORA workspace is an execution and collaboration platform driven by a community-based approach and is democratising access to advanced ICT solutions to accelerate the digitisation of manufacturing SMEs in an enriched environment. The Digital Agora was achieved in more than 7 years of European funded research and innovation actions –
        means, emGORA workspace has been nudged loose by the European Commission’s efforts to build up economic strengths in a sustainable way.
        To cover all the nations in Europe, emGORA workspace is joining forces with Digital Innovation Hubs (DIH) to offer economic opportunities for all in the manufacturing
        technology sector and beyond. emGORA workspace is aiming to play an essential part in the European Commission's sustainability strategy - for our Europe of tomorrow. The divers and open work environment respects and embraces differences, recognises the unique contributions that people with many types of differences can make, and most important
        maximises the potential of all those members who actively involve themselves.
        The newly created modular delivery-as-a-service platform combines the secure software marketplace powered with Cloud and HPC resources and stands out from others by offering
        a place for a human-centric community activities, which is focused on facilitating Pan-European collaborations.
        The marketplace of emGORA workspace focuses on the private and secure delivery of advanced software solutions for manufacturing and engineering by various providers and
        allows on the go execution, comparison, and testing of software tools, like CAx, Data analytics, simulations, AI and many more. Software services can be proprietary or open-
        source software and provided by any provider/company.
        The community area is all about networking and collaboration and aims to deliver a lot of the benefits from real manufacturing networks. SMEs, providers, and other members can interact with each other, explore tools and seek for help to tackle a certain challenge or digitise their production. The community is driven by exchange, every emGORA workspace
        member can host their own discussion roundtable, launch “Call for Solutions”, as well as to publish content (articles, success stories, use cases, insights, news) and share their
        perspective.
        Next to the delivery of software services via the marketplace, experts around advanced ICT solutions (providers and consultants) can propose their solution. They can challenge each other with the solution or can create joint forces to serve with the most suitable solution.

        By sustaining emGORA workspace as a result of publicly funded research in Europe, it offers the exchange of knowledge while protecting intellectual property rights of individuals.

        Speaker: Andreas Ocklenburg (cloudSME)
      • 57
        Q&A Buffer + Summary Wrap-up
        Speaker: Elisa Cauhe (EGI.eu)
    • Networking Reception in GatherTown https://go.egi.eu/EGI2021_NetworkingReception (GatherTown)

      https://go.egi.eu/EGI2021_NetworkingReception

      GatherTown

      Welcome to GatherTown!

      Look for your colleagues - if you're standing close enough, you can chat with each other!

      Looking for more privacy? Go sit at a table for 2 or with a small group.

      Room closes at 19.15
      Maximum capacity: 200 guests

    • 9:25 AM
      Chair Yoga go.egi.eu/egi2021-2 (Zoom Room 2)

      go.egi.eu/egi2021-2

      Zoom Room 2

    • 58
      70 years of discoveries with the Italian Institute for Nuclear Physics. Interview with Prof. Antonio Zoccoli, INFN President go.egi.eu/egi2021-1 (Zoom Room 1)

      go.egi.eu/egi2021-1

      Zoom Room 1

      Prof. Antonio Zoccoli will walk us through 70 years of discoveries in Physics with INFN, the Italian National Institute of High Energy Physics, member of the EGI Federation. With Antonio we will learn how innovation in ICT from broadband communications to distributed computing have been key enablers of scientific excellence.

      Speaker: Tiziana Ferrari (EGI.eu)
    • 59
      Keynote: Computing for Gravitational-wave research: now and tomorrow (Stefano Bagnasco, INFN) go.egi.eu/egi2021-1 (Zoom Room 1)

      go.egi.eu/egi2021-1

      Zoom Room 1

      Stefano Bagnasco is a senior Technology Researcher with the Istituto Nazionale di Fisica Nucleare in Torino, Italy. He has 20 years of experience in designing, maintaining and operating large-scale distributed infrastructures for scientific computing, first for HEP experiments and now in the field of Gravitational Waves research. He is currently serving as Computing and Data Processing Coordinator for the VIRGO experience and co-chair of the e-Infrastructure Board of the Einstein Telescope initiative.

      Speakers: Stefano Bagnasco (INFN), Tiziana Ferrari (EGI.eu)
    • 11:00 AM
      Break
    • Collaboration accross Boundaries - Workshop: Data Spaces and Data Ecosystems in Practice go.egi.eu/egi2021-2 (Zoom Room 2)

      go.egi.eu/egi2021-2

      Zoom Room 2

      Terms such as "data ecosystem" and "data space" are being used more and more in business and by policymakers, but it is unclear how these terms relate to the research world, where data sharing and exchange, and creating open data, are common objectives of most research projects. This session will define data ecosystems and data spaces by their key characteristics and explore the data exchange approaches being used in several international research communities. The session will explore how these research data initiatives map to the "data spaces" that have been proposed by the European Commission in a number of sectors and identify key challenges for the research community in responding to EC funding calls for the creation of data spaces.

      • IS-ENES (climate modelling data)
      • GBIF (biodiversity data)
      • SeaDataNet (ocean and marine data)

      Conveners: Andrea Manzi (EGI.eu), Mark Dietrich (EGI.eu)
    • Delivering Services and Solutions - Presentations go.egi.eu/egi2021-1 (Zoom Room 1)

      go.egi.eu/egi2021-1

      Zoom Room 1

      Convener: Ignacio Blanquer (UPVLC)
      • 64
        F-UJI : A Tool for the automated assessment and improvement of the FAIRness of Research Data

        Funders, publishers and scientific organizations have highly endorsed the adoption of FAIR principles (Findable, Accessible, Interoperable, and Reusable) to promote research data reusability and reproducibility.
        However, FAIR principles are high-level guidelines without explicit requirements for their implementation. Several aspects should be taken into account to translate the principles into practice. Practical solutions such as metrics and associated tools are required to support the assessment of FAIR compliance of research artefacts such as services and datasets. FAIR assessment tools are important for many stakeholders to assess the reusability of data and -related services and thus their suitability for supporting science. For the individual researcher as well as research initiatives these tools can help to select for example appropriate data archives, but they can also be used by data service providers to iteratively improve their data offerings, e.g., as part of FAIR advisory processes. The FAIRsFAIR (Fostering Fair Data Practices in Europe) project aims at providing practical solutions for the application of the FAIR data principles throughout the research data life​. One of the outcomes of the project is the development of an open-source tool named F-UJI to support trustworthy data repositories committed to FAIR data provision to programmatically measure datasets for their level of FAIRness over time. The tool supports a programmatic FAIR assessment of published research datasets based on the FAIRsFAIR object assessment metrics. For each of the metrics, we have designed and implemented practical tests based on existing standards and best practices for research data. This presentation gives an overview on the development and application of F-UJI and the underlying metrics and tests. We further demonstrate the use of the tool in a combined consultative approach applied to pilot data repositories. As part of iterative development of the tool, the automated assessment of multidisciplinary datasets was used to derive recommendations for improving the FAIRness of the datasets tested using the tool. We summarize the experience and lessons learned from the development, testing and application of F-UJI.

        Speaker bio:
        Robert Huber is a marine geologist and information specialist working in the PANGAEA working group at the Centre for Marine Environmental Sciences (MARUM, University of Bremen). Here he is responsible for projects in the fields of e.g. ontology development, marine observatory networks, scientific data management and FAIR data in general. In the context of the FAIRsFAIR project, Robert is leading Task 4.5 and responsible for the F-UJI FAIR assessment tool.

        Speaker: Robert Huber (Universität Bremen)
      • 65
        OIDC Support for Windows using Putty

        OpenID Connect (OIDC) is an identity layer on top of the OAuth 2.0 protocol. Relying on OIDC for identity and access management can significantly simplify the process of providing access to users, especially for non-web applications where the management of typically used SSH keys is often laborious and error-prone.
        This project aims to provide client-side support for OIDC access tokens for Putty - one of the most famous SSH clients for Windows. With the newly developed OIDC support for the server-side [https://github.com/EOSC-synergy/ssh-oidc], this effort fills the gap of the missing OIDC client functionality for Windows. Together, they provide a complete solution to manage users' access based on OIDC/OAuth2.

        The project consists of two parts. The first part aims to port the oidc-agent to Windows. Oidc-agent is a set of command-line tools for Linux and macOS that enables users to manage and obtain OIDC access tokens. It follows the design of the ssh-agent and, as such, it can be easily integrated into the user's flow. The expected impact of this subtask is potentially huge. Users from many communities that may use this tool (such as HPC or any other community that needs constant access to computing and storage resources as part of their work) typically rely on Windows as their primary operating system.

        In the second part of the project, we are going to integrate the oidc-agent with Putty. Users will be able to choose between using SSH with pageant - a Putty's ssh key manager or using SSH with OIDC-tokens against an OIDC-capable ssh-server.

        Speaker: Dmytro Dehtyarov
      • 66
        The EGI-ACE Cloud and HTC providers

        EGI-ACE federates together a distributed set of 16 hosting centres that bring compute and storage capacity for data processing, hosting research data and tools and scale up data analytics environments, scientific gateways and other added value services. Over the 30 months of the project, the hosting sites will provide more than 82 M CPU hours and 250 K GPU hours, as well as 45 PB/month through the Cloud Compute (IaaS) and High Throughput Compute services of the EGI portfolio. They will support EGI-ACE platforms and data spaces and those users coming from the use cases calls of the project and the EOSC portal. Alongside with the baseline computing and storage infrastructure, a set of enabling components support the Cloud Compute service: AppDB, for resource discovery and software catalogue; Dynamic DNS, for user-managed DNS provision of domain names for VMs and services running on the e-Infrastructure; and Infrastructure Manager (IM) , for the basic orchestration of IaaS resources. This presentation will provide an overview of the hosting centres and the enabling services. It will also showcase the main use cases supported in the infrastructure and the current usage of the available resources from the project.

        Speaker bio: Enol is Cloud Solutions Manager at the EGI Foundation working on the definition of the EGI Cloud federation architecture and its innovation roadmap. He also supports the EGI Community Team in the co-design and implementation of solutions for meeting the computing requirements of their research platforms and applications. Previously, he worked in the User Community Support Team of EGI and as middleware developer and providing support to user communities at UAB and CSIC in the context of distributed computing European projects. He holds a PhD in Computer Science from Universitat Autòmoma de Barcelona and a Computing Engineering Degree from Universidad de La Laguna.

        Speaker: Enol Fernandez (EGI.eu)
      • 67
        Connecting research data workflows with Argos: Embedding services and scientific outputs in data management plans

        Planning Research Data Management (RDM) activities is crucial as it effectively lays the foundations for successful research conduct according to Research, Technology and Development (RTD) best practices and requirements. Additionally, it leads to validated and quality research outcomes at the end of the scientific project / endeavour. Data Management Plans (DMPs) implement RDM policies and describe the activities, means and effort needed to manage data generated, collected and / or re-used by researchers. Moreover, they are seen as outputs that enable data reproducibility, continuously updated with new data as “living documents” of ongoing research. Hence, DMPs have rapidly entered Research Data Management Lifecycles (RDML), included at the beginning of the research process, and have been standardized in structure and content to be easily adoptable by research funders, institutions and communities.
        Following the scientific demand, unravels the necessity for flexible and action-capable tools to deal with data management plans that enable research links and information contextualization while also steering compliance with the FAIR data concept. Considering that an important factor to enabling FAIRness of data are the practices followed by research infrastructures and service providers, it poses the possibility for better communication and integrations among various service providers in support of diverse RDM settings (e.g. discipline-specific, multi- or inter-disciplinary, inter-institutional, cross-country and trans-national purposes). Research data services offer a complementary and fine-grained view to addressing concrete actions for the access, sharing and management of data handled as input and/or output in their workflows. Thus, services should become more prominent components of DMP outputs and clearly link to the FAIR and RDM aspects corresponding to their use by researchers.
        The presentation is about Argos, OpenAIRE’s machine actionable DMP service, also available as an EOSC (European Open Science Cloud) resource. Argos is a ready-to-use RDA compliant service for creating and publishing DMP outputs. It provides the scientific community with an open source, configurable and extensible instrument for the handling, validation and dissemination of their data management plans, in line with policies settled by a variety of funding agencies, institutions, research communities and other initiatives. Argos is hooked into the OpenAIRE ecosystem and interconnects with its underlying services and the OpenAIRE Research Graph to offer maximum capabilities for metadata contextualisation, integration with diverse research workflows, and validation of input data comprising DMPs. The presentation brings attention to the full DMP lifecycle supported by Argos and highlights collaborations, integrations and out-of-the-box links with OpenAIRE and EOSC services. Finally, it reflects on the role of DMPs from the perspective of service providers and communicates possibilities for embedding services in DMPs, while complementing and automating current RDML practices.

        Speaker bios:
        George Kakaletris https://gr.linkedin.com/in/gkakas

        Elli Papadopoulou has a BSc in Library Science and Information Systems. Master's in Public Policy and Public Administration (Athens University of Business and Economics). Special training in Research Data Management and work experience in academic and research environments: cataloging, classification, management of information systems, user requirements (University of Bath Library, United Kingdom and International Hellenic University Library), and management, preservation and archiving of digital documents (coordination of a digitization project at the CERN Library, Switzerland), For the past four years she has been working for ATHENA Research Center in Greece. In OpenAIRE, she undertakes the coordination of the National Open Access Desk in Greece (OpenAIRE Greek NOAD) along with the management of the Argos tool for the preparation and publishing of Data Management Plans. She also contributes to Open Science oriented projects at both national and european levels, such as the NI4OS-Europe (National Initiatives for Open Science in Europe) and HELIX (Hellenic Data Service). She has been volunteering at DOAJ since 2015 (associate editor) and at the RDA since 2016 (co-chair of the Early Career and Engagement Interest Group and of the Engaging Researchers with Data Management Interest Group). Her research activities revolve around Research Data Management, Open Science, Public Services and Policy Analysis.

        Speakers: Elli Papadopoulou (Athena Research & Innovation Center in Information Communication & Knowledge Technologies), Dr George Kakaletris (CITE)
      • 12:40 PM
        Q&A
    • Envisioning the Future - Presentations go.egi.eu/egi2021-4 (Zoom Room 4)

      go.egi.eu/egi2021-4

      Zoom Room 4

      Convener: Natalie Cernecka
      • 68
        Next-generation Green Computing

        Green Computing is dedicated to the reduction of energy and material consumption in information and communication technologies. Although serious energy-saving measures are put in place already today, they are being overtaken by accelerated advances of digitization.

        The steady rise of the Internet of Things (IoT) results in a drastic grow of the number of sensors. In the same way, the resolution of measuring devices continues to increase. As a result, not only industrial and scientific domains but also everyday life have to cope with ever-increasing data volumes that need to be processed and, thereby, lead to dramatic swells of the energy consumption. As published in Nature [1], the worldwide power consumption due to IT needs in 2030 will be approx. 20%.

        In our contribution, we will emphasize that substantial data reduction is indispensable to cope with future power demands. Sooner or later, everyone comes to the point that almost no raw data can be stored in the long term anymore, but only a comparatively tiny fraction of “relevant information” that needs to be extracted automatically from huge data streams by applying suitable machine learning methods.

        The observatory Square Kilometre Array (SKA) has to overcome all these challenges - its thousands of antennas will produce more data than the world-wide internet (see our last year’s contribution [2]). “Traditional” data compression is not sufficient, but it is a matter of extracting "relevant information” already during data acquisition in near-realtime. Due to the time constraint, information loss is inevitable and not reversible. To minimize the resulting “data irreversibility”, we propose a "Dynamic Life Cycle" (DLC) that extends existing Big Data life cycle models by introducing two feedback loops: one between the sensors and a nearby computing center, and one between world-wide distributed data archives and the sensors - both of which constantly optimize the sensors' control systems.

        DLC may contribute significantly to sustainability in compute ecosystems. Realizing DLC, however, is technically extremely demanding. Its core, the data reduction processes, is to be described in detail by metadata to ensure the FAIR principles and, ultimately, reproducibility of scientific results in view of non-existent raw data. Hereby, the metadata can be larger in volume than the archived data itself, since the permanently changing parameters of the sensors' controls and the states of the workflows for extracting the "relevant information" must be constantly recorded. The understanding how data irreversibility affects data reduction changes over time, meaning that the quality of archives must be steadily monitored (by comparison with simulation data). Searching huge data for "rare events” or “unknown signals" needs an efficient massively parallel computing, which is in its infancy in image processing and machine learning.

        Our presentation addresses the impact of DLC on Green Computing. New ways of working together are indicated for a machine learning-based “next-generation Green Computing”.

        [1] N. Jones: How to stop data centres from gobbling up the world’s electricity, Nature 561, 163 (2018).
        [2] H. Heßling, M. Kramer, S. Wagner: Data Challenges at the Square Kilometre Array (SKA), EGI Conference 2020.

        Speaker bio:
        Hermann Hessling studied Physics at the Universities of Münster, Goettingen, and Hamburg. He received the Ph.D. (Dr. rer. nat.) in Theoretical Physics and was appointed a postdoctoral research fellow at Deutsches Elektronen-Synchrotron (DESY) Hamburg (1993-1996). Since 2000 he has been Professor of Applied Informatics at the University of Applied Sciences (HTW) Berlin. His scientific interests include distributed high-performance computing and, in particular, extracting knowledge out of large-scale data in real-time.

        Speaker: Hermann Heßling (Verein für datenintensive Radioastronomie (VdR), and University of Applied Sciences (HTW) Berlin)
      • 69
        A high performance and real-time prototype for Anomaly Detection in a Datacenter

        The amount of companies and DataCenters that handle large volumes of data using Computer Information Systems are constantly growing. Besides, the computing infrastructures become more complex integrating different types of architecture (HPC, Cloud Computing, GPU, low-latency networking, etc). In addition, hundreds or thousands of users sharing the same resources make the systems more vulnerable. Whenever a security issue arise due to an internal or external access, the consequences can be very serious for both suppliers and the customers sharing resources. In the worst case, this may cause the partial or total loss of data. Furthermore, the cybercriminals use new techniques that require complex security systems with many services involved, which become more difficult for the system administrators to manage. The use of monitoring systems and its continuous update are intended for controlling the network traffic at real time as well as avoiding the complex and unknown threats.
        Intrusion Detection Systems (IDS) play an important role in a Datacenter with the purpose of generating alerts at the moment when a malicious event arises. By this way, the system administrator can immediately react to know what services are vulnerable. However, the system generates alerts based on known rules and it is not capable of detecting and classifying the anomaly traffic causing the administrator unaware of the attack. In recent research, the IDS are complemented with other tools and techniques of Machine Learning to allow the analysis and post-processing of those events. Supported by high-distributed and high-performance computing, the integration of those tools makes the processing of the data in real time as well as the detection of anomaly events as quickly as possible.
        This presentation will introduce the most used IDS as well as a description of our current deployed architecture based on Suricata. Apart from that, a high-distributed and high-performance computing prototype architecture will be presented, where different tools of fault-tolerant message systems ( like kafka) and processing systems like Spark Streaming, Hadoop and Data Lakehouses are integrated to collect the data from the IDS and transform it into a more efficient type of data automatically. This architecture will allow the post-processing of the traffic packets as well as the anomaly events using ML techniques to make a fast detection and correct classification of the anomaly events by the IDS.

        Speaker bio: https://www.linkedin.com/in/aida-palacio-hoz-633bb45a/?originalSubdomain=es

        Speaker: Ms Aida Palacio Hoz (IFCA)
      • 70
        CCP4 Cloud: Potential for a European-Wide Resource in Computational Crystallography

        Collaborative Computational Project Number 4 in Protein Crystallography (CCP4 UK), founded in 1979, has a mission to produce, maintain and distribute a world-leading, integrated suite of programs for the determination of biological macromolecular structures in 3d by means of X-ray crystallography and other biophysical techniques. Today, CCP4 Software Suite, used by estimated 25,000 researchers worldwide, represents an enabling technology and underpins academic and industrial research in structural biology on world-wide scale.

        Nowadays, experimental part of structural studies is represented by high-throughput lines at synchrotrons, producing tens of TBs of data on daily basis. The data are usually utilised in local research labs with own computational resources and software setups.

        Following general trends in modern computing, CCP4 have developed a framework for distributed data processing and structure solution in crystallography, the CCP4 Cloud. The development was released in 2019, and since that adopted by several crystallography centres in the UK and continental Europe. An instance of CCP4 Cloud, capable of satisfying UK’s needs in crystallographic computing, is maintained by CCP4 at STFC UK.

        The many advantages of online computing: centrally maintained software, databases and project data, geographically agnostic access to considerable computational resources, became particularly obvious during the COVID-19 pandemic. In this presentation, I will reflect on the advantages, desirability and feasibility of making a global, cloud-based resource for computational macromolecular crystallography in Europe, which could combine data streams from experimental facilities, software and computing services, and a uniform user experience based on an interactive, graphical, online access from all modern client platforms.

        Speaker bio:
        Dr Krissinel leads the Core Group of CCP4, which is responsible for the maintenance, development and distribution of the CCP4 Software Suite in protein crystallography. In 2000-2009, Dr Krissinel worked in the PDBe (former Macromolecular Structure Database, MSD) at the European Bioinformatics Institute, UK, where he became known to the structural biology community for his works in the field of protein-protein interactions (PPI) and protein structure analysis. His software for the PPI and oligomeric state analysis and prediction (PISA), and protein structure alignment (SSM/Gesamt) are widely used and regarded as de-facto standard tools in the field. Before his career in structural bioinformatics and crystallography, Dr Krissinel worked on the theory of time-resolved diffusion and spin control in radical reactions, quantum kinetics (1997-2000 Humboldt Fellow, University of Konstanz, Germany, 1989-1995 Russian Academy of Sciences), and molecular dynamics of metallic alloy microclusters (1995-1997 Argonne National Laboratory, USA).
        Weblinks: https://www.scd.stfc.ac.uk/Pages/Eugene-Krissinel.aspx

        Speaker: Dr Eugene Krissinel (STFC)
      • 12:30 PM
        Q&A
    • Innovating Services Together - Workshop: Initiatives for better evidence-based policies in the Public sector go.egi.eu/egi2021-3 (Zoom Room 3)

      go.egi.eu/egi2021-3

      Zoom Room 3

      Public authorities need innovative tools to develop better evidence-based policies. Citizens and Business should be involved in this process to better inform policy-making and to increase the trust in the authorities and the policies acceptance. Shared data, analytics and cloud improve policy making at all levels, national and local.
      The main objective of this session is to bring together the 5 projects (PolicyCloud, AI4PP, Decido, DUET, Intelcomp)funded by the DT-GOV-12 call to share their experiences in the public sector. We also aim to increase the collaboration between the EGI researchers and the public authorities in the decision making with the use of research data and advanced tools for the benefit of society.
      ●The PolicyCloud project exploits the potential of digitisation, big data and cloud to improve the modelling, creation and implementation of policies. Delivering a unique, integrated environment of datasets, data management, and analytic tools it addresses the full lifecycle of policy management in four thematic-areas (radicalisation, food-value chain, city environment, city services),leveraging the data management capabilities of the EOSC Initiative. The Project empowers the Citizens to contribute to data and policies related to their everyday-life. The onboarding of these solutions in the EOSC Portal offers a great opportunity to reach a wide audience.
      ●AI4PP (AI for Public Policy) is a joint effort of policy makers and Cloud/AI experts to unveil AI’s potential for automated,transparent and citizen-centric development of public policies. The project will deliver,validate and promote the AI4PublicPolicy Platform,offering innovative policy management on unique AI technologies. The AI4PublicPolicy Virtualized Policy Management Environment(VPME) integrated with EOSC facilitates access to the Cloud and HPC resources required to enable the project’s AI tools and to a wider use of the project’s developments.
      ●Decido project(eviDEnce and Cloud for more InformeD and effective pOlicies) aims to boost the use of EOSC by Public Authorities enabling innovation in the policy-making sector allowing cross-support and cross-collaboration,using secure compute and data intensive services. Decido involves citizens and local communities through co-creation activities for better targeted policies.
      ●DUET (Digital Urban European Twins) is a EU initiative which leverages the advanced capabilities of cloud, sensor data and analytics in Digital Twins,to develop more democratic and effective public sector decision-making.DUET Digital Twins provide virtual city replicas which simplifies the understanding of complex interrelation between traffic, air quality, noise and other urban factors. Powerful analytics predict the impacts of potential change to make better evidence-based operational decisions and long-term policy choices.
      ●Intelcomp develops a Competitive Intelligence Cloud/HPC Platform for AI-based Science, Technology and Innovation Policy-Making. Multi-disciplinary teams will co-develop analytics services, Natural Language Processing pipelines and AI workflows,exploiting EOSC open data and resources, HPC environments and federated operations at the EU, national and regional level. Ensuring a cooperative environment,different actors visualize, interact and analyze information. Through co-creation, IntelComp will adopt a living labs approach, engaging public policy makers, academia, industry, SMEs, local actors and citizens to explore, experiment with and evaluate STI policies. IntelComp is targeting domains aligned with the European Agenda and the Horizon Europe Missions:AI, Climate Change and Health.

      Conveners: Eleonora Testa (EGI), Elisa Cauhe (EGI.eu), Marieke Willems (Trust-IT)
      • 71
        Intro
        Speaker: Marieke Willems (Trust-IT)
      • 72
        Roadmap Data Driven Policy Cluster
        Speaker: Marieke Willems (Trust-IT)
      • 73
        PolicyCloud

        The PolicyCloud project exploits the potential of digitisation, big data and cloud to improve the modelling, creation and implementation of policies. Delivering a unique, integrated environment of datasets, data management, and analytic tools it addresses the full lifecycle of policy management in four thematic-areas (radicalisation, food-value chain, city environment, city services),leveraging the data management capabilities of the EOSC Initiative. The Project empowers the Citizens to contribute to data and policies related to their everyday-life. The onboarding of these solutions in the EOSC Portal offers a great opportunity to reach a wide audience.

        Speaker: Ricard Mune
      • 74
        Intelcomp

        Intelcomp develops a Competitive Intelligence Cloud/HPC Platform for AI-based Science, Technology and Innovation Policy-Making. Multi-disciplinary teams will co-develop analytics services, Natural Language Processing pipelines and AI workflows,exploiting EOSC open data and resources, HPC environments and federated operations at the EU, national and regional level. Ensuring a cooperative environment,different actors visualize, interact and analyze information. Through co-creation, IntelComp will adopt a living labs approach, engaging public policy makers, academia, industry, SMEs, local actors and citizens to explore, experiment with and evaluate STI policies. IntelComp is targeting domains aligned with the European Agenda and the Horizon Europe Missions:AI, Climate Change and Health.

        Speaker: Cecilia Cabello
      • 75
        DUET

        DUET (Digital Urban European Twins) is a EU initiative which leverages the advanced capabilities of cloud, sensor data and analytics in Digital Twins,to develop more democratic and effective public sector decision-making.DUET Digital Twins provide virtual city replicas which simplifies the understanding of complex interrelation between traffic, air quality, noise and other urban factors. Powerful analytics predict the impacts of potential change to make better evidence-based operational decisions and long-term policy choices.

        Speaker: Lieven Raes
      • 76
        DECIDO

        Decido project(eviDEnce and Cloud for more InformeD and effective pOlicies) aims to boost the use of EOSC by Public Authorities enabling innovation in the policy-making sector allowing cross-support and cross-collaboration,using secure compute and data intensive services. Decido involves citizens and local communities through co-creation activities for better targeted policies.

        Speaker: Antonio Filograna
      • 77
        AI4PublicPolicy

        AI4PP (AI for Public Policy) is a joint effort of policy makers and Cloud/AI experts to unveil AI’s potential for automated,transparent and citizen-centric development of public policies. The project will deliver,validate and promote the AI4PublicPolicy Platform,offering innovative policy management on unique AI technologies. The AI4PublicPolicy Virtualized Policy Management Environment(VPME) integrated with EOSC facilitates access to the Cloud and HPC resources required to enable the project’s AI tools and to a wider use of the project’s developments.

        Speaker: Alessandro Amicone
      • 78
        Panel: Connection to EOSC
        Speakers: Andrea Cristofori, Ron Dekker (CESSDA)
      • 79
        Wrap up
        Speaker: Marieke Willems (Trust-IT)
    • Demonstration: Using EGI Cloud infrastructure with fedcloudclient go.egi.eu/egi2021-4 (Zoom Room 4)

      go.egi.eu/egi2021-4

      Zoom Room 4

      • 80
        Demo: Using EGI Cloud infrastructure with fedcloudclient

        fedcloudclient is a command-line client and high-level Python package for interaction with EGI Federated Cloud. The client will allow users to perform the various OpenStack operations in EGI Federated Cloud. Executing any OpenStack command on any site with fedcloudclient is very simple: it requires only three options: site, VO and the command. For example:

        • Listing images in fedcloud.egi.eu VO on CYFRONET-CLOUD site:

        fedcloud openstack image list --vo fedcloud.egi.eu --site CYFRONET-CLOUD`

        • Listing all VMs in eosc-synergy.eu VO on all sites in EGI Federated Cloud:

        fedcloud openstack server list --vo eosc-synergy.eu --site ALL_SITES`

        Four modules are included in fedcloudclient:

        • fedcloudclient.checkin for operation with EGI Check-in like getting tokens,

        • fedcloudclient.endpoint for searching endpoints via GOCDB, getting unscoped/scoped token from OpenStack keystone,

        • fedcloudclient.sites manages site configurations

        • and finally fedcloudclient.openstack for performing OpenStack operations.

        The demonstration will provide tutorial of using fedcloudclient in EGI Cloud infrastructure, from simple examples to more complex scripting and programming with fedcloudclient.

        References

        • fedcloudclient reppository: https://github.com/tdviet/fedcloudclient

        • Documentation: https://fedcloudclient.fedcloud.eu/

        Speaker bio: Dr. Viet Tran is a senior researcher of the Institute of Informatics, Slovak Academy of Sciences (IISAS). His primary research fields are complex distributed information processing, grid and cloud computing, system deployment and security. He received M.Sc. degree in Informatics and Information Technology, Ph.D. degree in Applied Informatics from the Slovak University of Technology (STU) in Bratislava, Slovakia. He actively participates in preparations and solving a number of EU IST RTD 4th, 5th, 6th, 7th FP and EU~H2020 projects such as PROCESS, DEEP-HybridDataCloud, EOSC-Hub and EOSC-Synergy. He is the author or co-author of over 100 scientific publications.

        Speaker: Viet Tran (IISAS)
    • 12:45 PM
      Lunch Pimp your Badge!

      Pimp your Badge!

      https://indico.egi.eu/event/5464/page/94-pimp-your-badge
    • Demonstration: Using Dynamic DNS service in EGI Cloud infrastructure go.egi.eu/egi2021-4 (Zoom Room 4)

      go.egi.eu/egi2021-4

      Zoom Room 4

      • 81
        Using Dynamic DNS service in EGI Cloud infrastructure

        Nowadays, more and more services are dynamically deployed in Cloud environments. Usually, the services hosted on virtual machines in Cloud are accessible only via IP addresses or pre-configured hostnames given by the target Cloud providers, making it difficult to provide them with meaningful domain names.

        The Dynamic DNS service provides a unified, federation-wide Dynamic DNS support for VMs in EGI infrastructure. Users can register their chosen meaningful and memorable DNS host names in given domains (e.g. my-server.vo.fedcloud.eu) and assign to public IPs of their servers. By using Dynamic DNS, users can host services in EGI Cloud with their meaningful service names, can freely move VMs from sites to sites without modifying server/client configurations (federated approach), and can request valid server certificates in advance (critical for security).

        The tutorial will provide live demonstration and practical advice on using Dynamic DNS service in realistic user scenarios.

        References:

        • Dynamic DNS service portal: https://nsupdate.fedcloud.eu/
        • Documentation: https://docs.egi.eu/users/cloud-compute/dynamic-dns/

        Speaker bio:
        Dr. Viet Tran is a senior researcher of the Institute of Informatics, Slovak Academy of Sciences (IISAS). His primary research fields are complex distributed information processing, grid and cloud computing, system deployment and security. He received M.Sc. degree in Informatics and Information Technology, Ph.D. degree in Applied Informatics from the Slovak University of Technology (STU) in Bratislava, Slovakia. He actively participates in preparations and solving a number of EU IST RTD 4th, 5th, 6th, 7th FP and EU~H2020 projects such as PROCESS, DEEP-HybridDataCloud, EOSC-Hub and EOSC-Synergy. He is the author or co-author of over 100 scientific publications.

        Speaker: Viet Tran (IISAS)
    • 1:15 PM
      lunch
    • Demonstration: Using MATLAB on EGI for Open Science go.egi.eu/egi2021-4 (Zoom Room 4)

      go.egi.eu/egi2021-4

      Zoom Room 4

      • 82
        Using MATLAB on EGI for Open Science

        This demonstration will provide users with a step-by-step tutorial for accessing data via the EGI Data Services and analyzing it with MATLAB on the EGI’s Notebooks Service.

        Users can connect with their own MATLAB licenses to analyze available datasets on the EGI’s DataHub in the cloud on EGI resources. MATLAB live scripts allow users to build computational notebooks combining text, images, code and results which can be shared with their communities via the EGI cloud storage. To share research output between diverse user groups, users can call other languages (eg. Python) from MATLAB and save data in widely accessible, open formats.

        In this demonstration, attendees will learn how to
        • Access the EGI’s Notebook Service
        • Use their MATLAB licenses to run MATLAB on the EGI
        • Connect to the EGI DataHub to access publicly available datasets
        • Create a computational notebook containing images, text, code, and output, all in one document
        • Apply interactive controls for users to tune notebook parameters during runtime
        • Share their MATLAB code and data with other users using EGI cloud storage

        Speaker bio:
        Dr. Shubo Chakrabarti is the EMEA Science Gateway Strategist at MathWorks and helps researchers using and hosting online portals to effectively share and easily access MATLAB and Simulink for their research. Shubo earned his MSc at the Kings College London and his PhD in neuroscience at the Penn State University Medical College in the US. Before joining MathWorks, Shubo worked as a senior neuroscientist and project leader for several years at the Universities of Göttingen and Tübingen in Germany. He is an Alexander von Humboldt fellow and a reviewer for several scientific journals and the German Research Society (DFG).

        Yona Baskharoun: https://www.linkedin.com/in/yona-baskharoun/

        Nick Choi is a product manager at MathWorks focusing on integrations between MATLAB and a variety of popular online platforms.

        Speakers: Shubo Chakrabarti (MathWorks), Ms Yona Baskharoun (The MathWorks), Mr Nick Choi (The MathWorks)
    • Demonstration: Knowledge sharing and discovery across heterogeneous research infrastructures go.egi.eu/egi2021-4 (Zoom Room 4)

      go.egi.eu/egi2021-4

      Zoom Room 4

      • 83
        Knowledge sharing and discovery across heterogeneous research infrastructures

        Research infrastructures play an increasingly essential role in scientific research. They provide rich data sources for scientists, such as services and software packages, via catalog and virtual research environments. However, such research infrastructures are typically domain-specific and often not connected. Accordingly, researchers and practitioners face fundamental challenges introduced by fragmented knowledge from heterogeneous, autonomous sources with complicated and uncertain relations in particular research domains. Additionally, the exponential growth rate of knowledge in a specific domain surpasses human experts' ability to formalize and capture tacit and explicit knowledge efficiently. Thus, a knowledge management system is required to discover knowledge effectively, automate the knowledge acquisition based on artificial intelligence approaches, integrate the captured knowledge, and deliver consistent knowledge to agents, research communities, and end-users. In this study, we present a knowledge management system, called ENVRI-KMS, for ENVironmental Research Infrastructures, which are crucial pillars for environmental scientists in their quest for understanding and interpreting the complex Earth System.

        The ENVRI-KMS is a Knowledge-as-a-Service for ENVRI-FAIR research communities to document the development and operation processes of RIs and support them with their engineering and design decisions. In general, the ENVRI-KMS should (1) ingest technical results from ENVRIplus, FAIR assessment, the key sub-domains, and other tasks using a formal language for knowledge representation and proven semantic technologies; (2) provide services and tools to enable RI developers and data managers to browse, search, retrieve and compare RI technical statuses and technical solutions to development problems via available content; (3) provide content management tools for specialists in the ENVRI community to ingest new knowledge and control the quality of content; (4) also provide interfaces to other existing semantic resources, e.g., the service catalog of a future ENVRI-HUB to enhance knowledge discovery and cross-RI search, between knowledge services and the online presence of ENVRI resources.

        In the next phase, the development effort will mainly focus on the following aspects: (1) Continuous content ingestion and curation. The ENVRI-KMS team will improve the knowledge ingestion tool and continuously ingest the description (metadata) of high-quality results from the ENVRI community (e.g., sub-domain or RI developers), including development results (e.g., best practices, software technologies, recommendations, updated FAIRness assessment possibly generated by new tools) in the ENVRI-KMS, and make those descriptions FAIR for the community.
        (2) Continuous improvement of the ENVRI-KMS based on the feedback is received from the community. Extra features, e.g., for ENVRI-KMS discovery and recommendation, will be further explored. (3) The development and operation of the ENVRI-KMS will also follow the software engineering DevOps practices. The continuous testing, integration, and deployment pipeline will be established. (4) We will also extend the content maintenance to community specialists. In this way, we hope the community will play a key role in the ENVRI-KMS.

        Speaker bio's:
        Siamak Farshidi: https://siamakfarshidi.nl/
        Na Li: https://www.uva.nl/en/profile/l/i/n.li/n.li.html
        Zhiming Zhao is currently a senior researcher in the group of System and Network Engineering ( SNE) at University of Amsterdam ( UvA). He obtained his bachelor and master degrees in Computer Science from Nanjing Normal University ( NJNU) and East China Normal University ( ECNU) in 1993 and 1996 in China respectively. He obtained his Ph.D. in Computer Science from University of Amsterdam (UvA) in 2004. He has strong research interest in advanced computing and network technologies, time critical and data intensive systems, Cloud computing, scientific workflows and software agents. He coordinates research and development activities in the EU H2020 project SWITCH (Software Workbench for interactive time critical and highly self-adaptive cloud applications), and in the “Data for Science” theme in the EU H2020 environmental science cluster project ENVRIPlus. He also leads the research tasks of research sustainability in the EU H2020 VRE4EIC project, and of semantic linking in the EU FP7 ENVRI project.

        Speakers: Siamak Farshidi (University of Amsterdam), Ms Na Li (University of Amsterdam), Zhiming Zhao (EGI.eu)
    • Delivering Services and Solutions - Presentations: Data Management / Security go.egi.eu/egi2021-1 (Zoom Room 1)

      go.egi.eu/egi2021-1

      Zoom Room 1

      This session will cover on the first part presentations on Data Management focusing on Storage solutions in the EGI infrastructure, their status and plans. In particular we will cover the status of DPM, with the support coming to an end and experiences implementing migration tools, and EOS which is currently under integration in EGI.
      The second part of the session includes two security presentations, covering the evolution of the Resource Trust and the EGI Software Vulnerability Group ( SVG ) activities and plans

      Convener: Andrea Manzi (EGI.eu)
      • 84
        DPM status and migration tools
        Speaker: Mr Peter Vokac
      • 85
        dCache evolution
        Speaker: Tigran Mkrtchyan (DESY)
      • 86
        EOS Open Storage for Science
        Speaker: Luca Mascetti (CERN)
      • 87
        StoRM
        Speaker: Enrico Vianello (INFN)
      • 88
        Resource Trust Evolution and Cloud-based Provisioning

        Increasing use of cloud resources, and other developments in new
        workflows, has raised the question of which providers of host
        certificates are appropriate for different use cases - with particular
        focus on the provisioning of dynamic compute resources. Any discussion
        of updating our trust model to include the use of new certificate
        authorities must include the perspectives of all stakeholders, including
        experiments and operations, sites, identity management and security. We
        present here a summary of the challenge facing our community and the
        current status of work to clarify the needs of our stakeholders and
        identify a path forward that meets our requirements

        Speaker: David Kelsey (STFC)
      • 89
        The EGI Software Vulnerability Group (SVG) - what we do and how we are evolving.

        The main purpose of the EGI Software Vulnerability Group can be summarized as "To minimize the risk of security incidents due to software vulnerabilities."

        The EGI SVG and its predecessors have been active since 2006, helping sites and VOs avoid incidents due to software vulnerabilities across the distributed computing infrastructure in a consistent manner.
        Initially the focus was on Grid computing and on software enabling Grid computing largely developed by our collaborators, but SVG has been handling a wider range of software which is used across the EGI infrastructure. Most of the work has been carried out by the SVG Risk Assessment Team (RAT), who assess the relevance of any reported vulnerability and the risk to the infrastructure. If a vulnerability has not been fixed yet, the RAT then asks the software provider to fix on a timescale according to the risk.

        In recent years the EGI infrastructure has become much less homogenous, there has been a proliferation of software used, and it is no longer possible for the SVG RAT to be experts in all the software used and how it is set up. So we are having to evolve to cope just like the world is evolving.

        This talk will describe the basic vulnerability handling procedure, and how the procedure is gradually being extended to deal with the less homogenous EGI infrastructure. This will include the concept of the "Deployment Expert Group" or "DEG", which consists of people who either run relevant services or are involved in the design or configuration of such services, to help handle vulnerabilities in this much less homogenous infrastructure. We will describe progress on evolving this procedure to date, and plans for the coming months and beyond. In particular, the DEG is open for more people to join and contribute to the continued security of EGI!

        Speaker: Linda Cornwall (STFC)
    • Delivering Services and Solutions: EGI Federation Tools and Services: Roadmaps for EGI and EOSC go.egi.eu/egi2021-2 (Zoom Room 2)

      go.egi.eu/egi2021-2

      Zoom Room 2

      This session will present the status and technical roadmaps of the Federation and Collaboration Tools as developed within the EGI-ACE project. This will cover new functionality that in the pipeline to address new requirements from users and other stakeholders.

      Conveners: Alessandro Paolini (EGI.eu), Matthew Viljoen (EGI.eu)
    • Envisioning the Future - Presentations: AI for the Common Good go.egi.eu/egi2021-4 (Zoom Room 4)

      go.egi.eu/egi2021-4

      Zoom Room 4

      Convener: Ville Tenhunen
      • 97
        AI experiences on EGI

        Nowadays, artificial intelligence (AI), machine learning, deep learning and data science are an essential part of the scientific research toolbox in most disciplines. These are also present in a number of projects, research initiatives and infrastructures. These new tools are also changing the traditional research ICT landscape. The EGI Federation as an international infrastructure - that provides advanced computing and data analytics services for research and innovation - has also faced these challenges and winds of change in the landscape.

        In this short presentation we describe AI experiences on EGI to the research infrastructure landscape and project landscape. How have these evolved and what are next steps?

        Speaker: Ville Tenhunen
      • 98
        StairwAI - A new service layer

        Aims

        The StairwAI project aims to create a bridge between users in a low-tech level to the higher-level AI resources. The project will do this by facilitating low-tech users' engagement on the AI on-demand Platform. This will be achieved through a new service layer enriching the functionalities of the on-demand platform and containing:

        (1) a multi-lingual interaction layer enabling conversations with the Platform in the user’s own language,

        (2) a horizontal matchmaking service for the automatic discovery of AI assets (tools, data sets, AI experts, consultants, papers, courses etc.) meeting the user requirements and,

        (3) a vertical matchmaking service that will dimension and provision hardware resources through a proper hardware provider (HPC, Cloud and Edge infrastructures).

        Objectives of the StairwAI project

        Societal:

        • To provide a framework compliant with the principles of ethical AI
          use

        Technological:

        • To ease the access to the AI-on-demand platform through natural
          language interaction
        • To satisfy the end user needs through horizontal matchmaking services
        • To dimension the physical resources of the AI on demand platform with
          end user requirements

        Economic:

        • To contribute to strengthen the European SMEs – including low tech
          SMEs – and DIH
        • To promote the sustainability of the AI on-demand platform
        • To reduce fragmentation of AI research and development

        AI techniques in use

        • Natural language processing (NLP) in different languages for easing
          low-tech users’ interaction
        • Knowledge representation for organizing the platform AI assets, reputation and fairness mechanisms
        • Constraint solving, optimization and machine learning to satisfy users’ business needs

        Open calls

        60 SMEs in low-tech sectors will be funded.

        Pilot calls:

        28 low tech SMEs will be selected in two different Open Calls: StaiwAI will support the preparation of the feasibility plan for the adoption of AI solutions by the low-tech sector SMEs and delivering a pilot for the adoption of AI.

        Call for adopters:

        32 low tech SMEs will be selected in one Open Call: StairwAI will support just the feasibility plan for the adoption of previously piloted AI solutions by low-tech sector SMEs.

        (This project has received funding from the European Union’s Horizon 2020
        research and innovation programme under grant agreement No 101017142)

        Speakers: Prof. Michela Milano (University of Bologna), Dr Gabriel G. Castañe (University College Cork)
      • 99
        AI4PublicPolicy will bring AI to support decision making

        AI4PublicPolicy is a joint effort of policy makers and Cloud/AI experts to unveil AI's potential for automated, transparent and citizen-centric development of public policies. To this end, the project will deliver, validate, demonstrate and promote a novel Open Cloud platform (i.e. AI4PublicPolicy platform) for automated, scalable, transparent and citizen-centric policy management based on unique AI technologies.

        The AI4PublicPolicy platform will be an Open Virtualized Policy Management Environment (VPME) that will provide fully-fledged policy development/management functionalities based on AI technologies such as Machine Learning (ML), Deep Learning (DL), NLP and chatbots, while leveraging citizens' participation and feedback. It will support the entire policy development lifecycle, based on technologies for the extraction, simulation, evaluation and optimization of interoperable and reusable public policies, with emphasis on citizen-centric policies development and optimization through the realization of citizen-oriented feedback loops.

        AI4PublicPolicy will complement public policy development functionalities with the ever-important process re-engineering and organization transformation activities towards ensuring the effective transition from legacy policy development models to emerging AI-based policy making.

        The AI4PublicPolicy VPME will be integrated with European Open Science Cloud (EOSC) with a dual objective. First to facilitate access to the Cloud and High Performance Computing (HPC) resources of EOSC/EGI that are required to enable the project's AI tools, second to boost the sustainability and wider use of the project's developments.

        AI4PublicPolicy's business plan for sustaining, expanding and commercializing the AI tools and the VPME is based on the development of a community of interested and engaged stakeholders (i.e. public authorities and other policy makers) around the project's platform.

        In this presentation we focus on the idea of the Open Virtualized Policy Management Environment (VPME), its' architecture and solutions.

        Speakers: Alessandro Amicone (GFT), Dr Ioannis Christou (Intrasoft), Dr Andrea Cristofori (EGI Foundation )
      • 100
        LETHE - A personalized prediction and intervention model for early detection and reduction of risk factors causing dementia, based on AI and distributed Machine Learning

        Aim of the LETHE project is to create a personalized prediction and intervention model for early detection and reduction of risk factors causing dementia, based on artificial intelligence (AI) and distributed machine learning (ML).

        The roots of LETHE project are deep and based on the result of research, studies and clinical data collected over the past 10 years. Memento, PredictND and the FINGER Study (Finnish Geriatric Intervention Study to Prevent Cognitive Impairment and Disability) are among the past projects and studies whose knowledge have become an integral part of LETHE project over time.

        Thanks to the heritage of the FINGER study, now widespread worldwide, it has been possible to demonstrate that prevention of cognitive decline in individuals at risk, by intervening on various aspects of lifestyle, is an effective health strategy. The LETHE project sets out to evolve the results of this impressive clinical study into a digital model - FINGER - based on the Internet of things, Mobile, Big Data and Artificial Intelligence (AI) technologies. The strategies for the prevention of cognitive decline will therefore be directed towards technological horizons usable and accessible to the ageing European population.

        LETHE will establish novel digital biomarkers, for early detection of risk factors, based on unobtrusive ICT-based passive and active monitoring. The aim is to establish a digital-enabled intervention for cognitive decline prevention based on the evolution of a successful protocol (FINGER) evolving into an ICT-based preventive lifestyle intervention through individualized profiling, personalized recommendations, feedback and support (FINGER), well targeted on a population stratified by cost-effective biological biomarkers.

        The LETHE solution will be tested in a feasibility study validating the achieved improvements. Upon success LETHE will provide individual-related, specific biomarkers enabling a more personalized risk factor prevention for persons with beginning cognitive decline, thereby empowering people to an active and healthy lifestyle. Expansions of prevention trials on large scale by an a multimodal intervention approach, reaching out to large populations, could save future costs on expensive traditional interventions and confer benefits for the wider society.

        In the LETHE project 15 partners from 9 European countries build a platform to store, integrate, analyse and share this sensitive data by nature within the consortium. Data sensitivity affects to all parts of the data pipeline of this platform like data ingestion, unstructured and structured data storing, data processing, AI/ML model processing, AI/ML OPS and data presentation. The LETHE project will also integrate tools and services released by European Open Science Cloud (EOSC).

        In this presentation we focus on conceptual architecture of the LETHE, AI solutions as a part of it and secure data management.

        Speakers: Ville Tenhunen, Mr Catalin Condurache (EGI Foundation), Mr Roberto Pasciuti (Extra Red), Monica Balatresi (Extra Red), Dr Carlo Aliprandi (Extra Red)
      • 101
        AI4EU Funding opportunities for AI experts and infrastructure providers

        Europe's AI4EU community brings together all individuals and organisations that are interested in using or benefitting from today's AI capabilities. Researchers, developers, entrepreneurs, citizens, civil servants and AI experts are invited to engage with the AI Ecosystem stakeholders and access multiple resources and solutions.

        Some of the ongoing projects under the AI4EU initiative are using the Financial Support to Third Parties (FSTP) scheme to identify disruptive bottom-up experiments contributing to the adoption of AI by SMEs. To this end, different Open Calls are going to be launched in the upcoming months and at the same time, parallel Open Calls for providers will onboard infrastructure and AI experts to support such experiments.

        The main purpose of this session is to share the AI4EU funding opportunities for the EGI community, mainly as infrastructure providers and AI experts to contribute and support the AI adoption by SMEs. The description of the objectives, the technical requirements, the funding conditions, the expected contribution and the application process will be presented in this session followed by a Q&A space to solve the doubts from the audience.

        It is expected that AI experts and EGI infrastructure providers willing to work with SMEs will participate in this session.

        Speakers: Elisa Cauhe (EGI.eu), Izabela Zrazinska
    • 3:30 PM
      Break
    • Poster Session go.egi.eu/egi2021-4 (Zoom Room 4)

      go.egi.eu/egi2021-4

      Zoom Room 4

      Convener: Gwen Franck
      • 102
        • Building a prototype for Earth Science Data and HPC Services

        In today’s world, more and more data are constantly being generated, resulting in changing the nature of computing, with an increasing number of data-intensive critical applications.

        As the Council adopted a regulation on establishing the European High Performance Computing Joint Undertaking (EuroHPC) in July 2021, the regulation paves the way for the development within Europe of the next generation of supercomputers, strengthens research and innovation capabilities, the development of a supercomputing infrastructure ecosystem and the acquisition of world-class supercomputers.

        PHIDIAS (Prototype of HPC /Data Infrastructure for On-demand Services) will be building a prototype for Data/High Performance Computing services based on Earth sciences cases. The aim of the project is to enable the Earth science community to discover, manage and process spatial and environmental data spanning the Earth’s surface, the atmosphere, and oceans. The project foresees the development of three use cases.

        • Intelligent screening of a large amount of satellite data for detection and identification of anomalous atmospheric composition events;
        • Processing on-demand services for environmental monitoring; and
        • Improving the use of cloud services for marine data management.

        In addition to developing datasets to be added into the EOSC catalogues, PHIDIAS will optimise workflows to facilitate data reuse, will provide open access to standardised HPC services, and will render the data FAIR.

        PHIDIAS will explore a distributed model for data transfer and resource allocation between two European computing centres (CINES in France and CSC in Finland). The project will develop data post-processing methods coupled with HPC capabilities, which will be deployed as a service for several end-users (scientific communities/public authorities/private players/citizen scientists).

        The data generated and services created will be available on the relevant EU portals (EU Open Data Portal/EUDAT/EOSC) and will be preserved using the long-term preservation services of the EOSC. This PHIDIAS project will then propose a generic workflow for massive scientific data by combining computing, dissemination and archiving resources in a single framework.

        Speakers: Mr Francesco Osimanti (Trust-IT Services), Julie Arteza (Trust-IT Services)
      • 103
        Making data-driven policy management a reality across Europe

        Digitisation and big data are revolutionising the way we live, think and make decisions. With the constant proliferation of devices and networks, more and more of the data being created from a variety of sources is able to be stored, transmitted and processed efficiently in large amounts. Increasingly, cloud environments, big data and other innovative data-driven approaches are enabling policy makers to deliver evidence-based policies which impact economic growth and contribute to modernising the public sector and improving the lives of citizens across Europe.

        However, new technologies and big data also bring new challenges. With so much data now available to policymakers, the search is on for ways to effectively access, make sense of, and use data to model, create and implement policy successfully. The Policy Cloud project addresses these challenges by using the European cloud infrastructure for public administrations. The 15-partner-strong consortium will deliver a unique integrated environment addressing the full lifecycle of policy management. Policy Cloud will use the capabilities offered by the European Open Science Cloud Initiative, especially concerning data analysis, in order to facilitate evidence-based policy making.

        Policy Cloud runs four pilot cases in different fields which will serve as demonstrators for data-driven policy management and leverage methodologies for user participation from citizens and communities. The results will be evaluated in heterogeneous environments and settings, demonstrating the applicability and re-usability of the developed models and tools in different countries, contexts, and sectors.

        1. Policies against radicalisation: will focus on collecting and analysing data from social media and message boards, in order to provide policy makers with effective data-driven tools for addressing the issue of radicalisation.
        2. Intelligent policies for the food value chain: will exploit the Policy Cloud platform as a support tool to facilitate collaborative decision making for the various actors in the agrifood value chain in Aragon, Spain.
        3. Urban policy making through analysis of crowdsourced data: the Municipality of Sofia will seek to improve urban environments by crowdsourcing data analysis, specifically via an interactive map of the city on which citizens can signal issues to the administration in real-time.
        4. Open Data Policies for Citizens: will build upon the work initiated in London on predictive analysis to identify unemployment risk.

        The final solutions and tools developed by Policy Cloud will eventually become available as public cloud services, contributing to the development of the European Open Science Cloud (EOSC), a core element in the European Commission's vision of making Europe a global leader in data management. Policy Cloud’s work for the registration of the cloud-based environment for data-driven policy management in EOSC, as lead by EGI, will provide the grounds for the provisioning of the required EOSC cloud environments and services to policy makers in the scope of the integrated Policy Cloud platform.

        Policy Cloud has received funding from the European Union's Horizon 2020 research and innovation programme under grant agreement No. 870675.

        Speakers: Marieke Willems (Trust-IT), Zach Smith (Trust-IT)
      • 104
        Enhancing collaborative ocean science with the Blue-Cloud services

        Blue-Cloud is the thematic EOSC for the marine domain supporting FAIR and Open Science by developing a web-based environment to provide scientists & researchers with enhanced analytical capabilities and cloud-computing resources, underpinned by simplified access to a wealth of multidisciplinary and interoperable marine data services.
        The project builds on existing European marine data infrastructures and e-infrastructures federating their services within the Blue-Cloud framework, enabling researchers to find, access, share, combine and reuse quality data across disciplines and countries. These infrastructures include key players such as EMODnet, Copernicus, SeaDataNet, Euro-Argo, D4Science, EUDAT, and more.
        The federation is taking place at the levels of (meta)data resources, computing resources and analytical service resources, driven by collaboration across research, data and e-infrastructures.

        • A Blue-Cloud Data Discovery and Access Service (DDAS) is being finalised to facilitate access to multi-disciplinary datasets. The overall concept is that the DDAS functions as a broker both for metadata and for data access, interacting with web services and APIs from each of the blue data infrastructures (BDIs) as federated in the Blue-Cloud. This way, it facilitates users to discover first at the collection level which BDIs might have data sets, interesting for their use case, and next, to identify and download relevant data sets at granule level from those selected BDIs, by means of a common interface.
        • A Blue Cloud-Virtual Research Environment (VRE) has been established to enable collaborative research. Services include Data Analytics (Data Miner, Software and Algorithms Importer (SAI), RStudio, JupyterHub), Spatial Data Infrastructure to store, discover, access, and manage vectorial and raster georeferenced datasets, and services and components enabling users to document and then either share with selected colleagues or make available online any generated product (e.g. analytical methods, workflows, processes, notebooks).
        • This innovation potential is explored and unlocked by a series of Virtual Labs developed by five real-life demonstrators, addressing societal challenges in the domains of genomics, fishery, aquaculture, biodiversity and environment.

        The demonstrators are showcasing how Blue-Cloud can enhance collaborative research in support of the EU Green Deal and key international initiatives, such as the United Nations Decade of Ocean Science for Sustainable Development.
        The poster highlights the key services developed within the Blue-Cloud technical framework and their potential impact on marine research towards tackling global challenges.

        Federico Drago is a Copywriter and Digital Marketing Specialist, currently involved in projects related to the European Open Science Cloud (EOSC) and to ocean sustainability. Holding a Master’s degree in English and Chinese from the University of Naples “L’Orientale”, I have worked in a range of international environments between Italy and China, including education, tourism, and culture at large. Outside of Trust-IT, I write about metal music & the climate crisis.

        Speaker: Federico Drago (Trust-IT Services)
      • 105
        ScienceMesh – Federated Collaboration Platform for Researchers

        The usage of EFSS (Enterprise File Sync&Share) platforms for sharing files is increasing, with a global market expected to reach $17 billion by 2025. Within the research and education sector, the Cloud Storage Services for Synchronisation and Sharing (CS3) community brings together vendors, service providers and users of EFSS platforms. CS3 provides services that are indispensable for the daily user workflows, allowing researchers, scientists and engineers to backup, share, transfer and synchronise data in seamless yet powerful ways. However, these services are largely disconnected and deployed in isolation from one another and other research services, preventing an effective global collaboration and exploitation of data towards scientific and economic progress. Users need to be able to globally share and collaborate on datasets, as well as use higher-level services, such as computing services for data analysis. This is where Science Mesh comes into play.
        Science Mesh will provide the aforementioned users with an interoperable platform with data, applications and computation combined, enabling them to easily synchronise, share and collaborate through applications and software components across Mesh-powered sites.
        Science Mesh allows the best of both worlds: users do not need to leave the well-known interfaces of their institutional services to be able to efficiently collaborate with users in other institutions. Its unique functionalities may be easily customised to the needs of particular research disciplines. And it is leveraging a fully-Open-Source development model in close collaboration with the Open-Source software industry while improving the Technology Readiness Level of contributing technologies (e.g. OpenCloudMesh). Another added value is the possibility for users to contribute in developing application plugins or easily set-up federations in their own scope and develop their own activities on top of them.
        Science Mesh users (researchers, educators, data curators and analysts) will gain the ability to share their datasets widely according to FAIR principles and respecting the GDPR , without losing control over them. It will be an integral part of the European Open Science Cloud (EOSC), offering researchers opportunities to assemble an efficient, reliable, collaborative and transparent research toolchain.
        Bob Jones, Director of EOSC Association adds, “The project has the potential to deliver a collaborative cloud-based data-sharing service for Europe, linking different communities and enabling cross-disciplinary research.” Science Mesh is being developed in close contact with pilot research communities, including Earth Observation (JRC), High Energy Physics (LHC), Astronomy (LOFAR) and Cultural Heritage Studies (PARADISEC). Science Mesh can reach a wide audience of stakeholders interested in collaborating for increasing scientific knowledge, coming from both the academia and the research industry.
        Future market opportunities for its commercial use are under study with global IT service companies (e.g Ailleron/SoftwareMind). The federated testbed already exists, connecting eight initial sites.

        Speaker bio: https://www.linkedin.com/in/ritasofiameneses/

        Speaker: rita sofia meneses (trust-it services)
      • 106
        The future of Open Science asks for a common understanding

        Abstract:
        The goal of this poster is to contribute to a common understanding of the concept of "Open Science".

        Indeed, it is widely recognized nowadays that there is no single, accepted, unified definition of Open Science. This motivates our proposal of an Open Science definition as a political and legal framework where research outputs are shared and disseminated in order to be rendered visible, accessible, reusable.

        Our contribution is developed standing over the concepts enhanced by the Budapest Open Science Initiative (BOAI), and by the Free/Open
        Source Software (FOSS) and Open data movements. We have elaborated this
        definition proposal through a detailed analysis of some selected EC policies and laws as well as of the function of research evaluation practices. The legal aspects considered in our examination include, in particular, the study of the role of licenses in the context of the dissemination of research outputs.

        This poster summarizes the content of our recent paper entitled: "Towards an Open Science definition as a political and legal framework: on the sharing and dissemination of research outputs" Teresa Gomez-Diaz, Tomas Recio,
        Version 2 published on POLIS N. 19, 2020,
        http://uet.edu.al/polis/images/Teresa_Gomez-Diaz.pdf,
        and Version 3 dated 28/02/2021 with minor corrections available on Zenodo, https://zenodo.org/record/4577066

        T. Gomez-Diaz is a CNRS Research Engineer at the Gaspard-Monge Computer Science laboratory (LIGM) at the University Gustave Eiffel (Est of Paris), where her mission is, since 2006, to render the research software production visible and accessible. This mission has been recently extended to an Open Science mission.

        She has been involved on the PLUME Project (2007–2013), where she has participated to launch and build a national catalogue of (French) research software. She has also participate to propose Software Management Plan tools. In collaboration with Prof T. Recio (University of Nebrija, Spain), she has proposed the CDUR research software evaluation procedure and an Open Science definition as the political and legal framework where research outputs are shared and disseminated in order to be rendered visible, accessible and reusable.

        Speaker: Teresa Gomez-Diaz (CNRS)
    • Delivering Services and Solutions - Workshop: EGI Data Transfer WG activities go.egi.eu/egi2021-3 (Zoom Room 3)

      go.egi.eu/egi2021-3

      Zoom Room 3

      EGI launched a ‘Data Transfer Working Group’ to drive the technical evolution of Data Transfer services in the context of the EGI federation.
      This workshop is organised to engage with scientific communities and technology/services providers of the domain to present use cases, pilots, state-of-the-art solutions.
      The workshop is relevant for scientific users, scientific communities who need to transfer large amounts of data among institutes at national, international or inter-continental scale. The session is relevant for developers and operators of data transfer services who want to propose solutions for EGI providers and user communities.

      Convener: Andrea Manzi (EGI.eu)
      • 107
        EGI Data Transfer WG session intro
        Speaker: Andrea Manzi (EGI.eu)
      • 108
        Data Transfer in the Science Mesh - CS3MESH4EOSC
        Speakers: Antoon Prins, Ron Trompert (SURFsara BV)
      • 109
        Sharing Data while Keeping Data Ownership

        The session is an introduction to the EUHubs4data project (https://euhubs4data.eu/), and presents one of the challenges data owners face when publishing their datasets, by being unable to safeguard, share or monetize their data. The session covers the solutions adopted in the project to address this challenge, and how an architecture based on the IDSA ecosystem empowers Digital Innovation Hubs (DIHs) to achieve a FAIR usage of their datasets. A short demo about how data from EGI DataHub can be published through an IDSA connector rounds up the session.

        Speaker: Levente Farkas
      • 110
        FTS development updates
        Speaker: Mihai Patrascoiu (CERN)
      • 111
        HIFIS transfer service
        Speaker: Tim Wetzel
      • 112
        What’s New in Globus
        Speaker: Vas Vasiliadis (University of Chicago)
    • Innovating Services Together: Presentations go.egi.eu/egi2021-1 (Zoom Room 1)

      go.egi.eu/egi2021-1

      Zoom Room 1

      Convener: Valeria Ardizzone (EGI.eu)
      • 113
        WORSICA, a Water Monitoring Cloud Platform using Sentinel Imagery.

        WORSICA (Water Monitoring Sentinel Cloud Platform) is a one-stop-shop service that integrates remote sensing (Copernicus) and in-situ data for the determination of water presence in coastal and inland areas, applicable to a range of purposes from the determination of flooded areas (from rainfall, storms, hurricanes or tsunamis) to the detection of large water leaks in major water distribution networks. It is one of the thematic services developed under the EOSC-Synergy initiative.
        WORSICA provides access to customized remote sensing applications currently applied to three major water sub-services: Coastline Detection, Inland Water Bodies Detection, and Leak Detection in Irrigation networks.
        In the Coastline Detection application, the user will obtain the coastline determined from Sentinel-2, UAV, or Pleiades imagery, for a given Region-Of-Interest. In the future, the coastline detection will be linked to the OPENCoastS service in order to produce a Digital Elevation Model. In the Inland Water Bodies Detection, the user will obtain layers of water bodies detected in inland regions (e.g., lakes or reservoirs) to characterize their volume and occupied/inundated areas. The Leak Detection in Irrigation networks application, the main goal of this sub-service, can automatically identify and detect possible water leaks in (remote) irrigation networks, obtained systematically from satellite and UAVs images.
        This service is aimed at Partners, Companies, and the Scientific Community. It can provide tools to the Partners (e.g., coastal engineers, researchers, and water supply managers) to monitor and anticipate the impacts of eventual storm surges/inundations on the coast and water leaks in irrigation networks. In addition, it can provide all Companies (e.g., coastal managers) the historical morphology of the coastline for specific periods in time, supporting multiple uses such as water monitoring harbor activities, dredging works, and building works on the coast. Finally, these detections are also helpful for the Scientific Community, supporting, for instance, fieldwork and helping to understand the evolution of the coastline and erosion patterns in the coastal areas and other water systems.
        WORSICA is a remarkably complex and computationally demanding service with several technological needs, from GRID and Cloud computing, storage and GPUs. As the service will serve the entire European scientific community, it must be efficient, scalable, and resilient. Thus, new automatic integration methodologies in the marketplace catalog and federated computational infrastructures of the EOSC network were implemented in the new architecture of the WORSICA service. WORSICA will use the available HPC computation resources of the EGI infrastructure to provide speedup, scalability, portability, flexibility, redundancy, and service interoperability. In addition, WORSICA takes advantage of several essential EGI services such as EGI CheckIn for federated authentication across IT services and infrastructures, Software Quality as a Service (SQAaaS) implemented using Jenkins pipelines for CI/CD in service integration automation, and Dataverse for logging metadata of processed products for FAIRness data.
        This work will present the main architecture changes implemented in the scope of the EOSC-Synergy project to the WORSICA service.

        Speaker info:
        http://www.lnec.pt/hidraulica-ambiente/en/team/ricardo-jose-da-silva-martins/

        Speaker: Ricardo Martins (Laboratório Nacional de Engenharia Civil)
      • 114
        CloudBank EU

        The vast amounts of data generated by scientific research pose enormous challenges for capturing, managing and processing this data. Many trials have been made in different projects (such as HNSciCloud and OCRE), but today, commercial cloud services do not yet play a major role in the production computing environments of the publicly funded research sector in Europe. Funded by the Next Generation Internet programme (NGI-Atlantic) from the EC, in partnership with the University California San Diego (UCSD), CERN is piloting the use of CloudBank in Europe. CloudBank has been developed by the UCSD, University of Washington and University of California, Berkeley with NSF grant support, to provide a set of managed services simplifying access to multi-cloud services for research and education, specialised in cost management and optimisation, that supports diversification of sources of funding, in a scalable “bring your own contract” model, across diverse research projects and multiple organisations. The European NGI experiment is provisioning cloud services from multiple vendors and deploying a series of use-cases in the domain of Machine Learning, HPCaaS, QCaaS and DBaaS, contributing to the scientific programme of the Large Hadron Collider. The main objective is to address technical, financial and legal challenges to determine whether CloudBank can be successfully used by Europe’s research community as part of its global research activity.

        Speaker bio:
        Apostolos' background lies at the intersection between computer science, data science, statistics and management. He has recently completed his master program in Business Analytics in University of Geneva while he also holds a bachelor in Computer Science. He has previously worked in the software industry in Greece and Italy as well as in International Organizations in Geneva on data processing, analysis and web development. He is currently working at CERN’s IT department on data exportation, processing, analysis and visualization to provide metrics and analytics for optimizing processes and costs.

        Speaker: Apostolos Theodoridis
      • 115
        Federated access for SSH with OpenID Connect

        Federated Identity Management, as modelled in the AARC blueprint architecture, has established itself as a de-facto standard for authentication and authorisation in research infrastructures.

        Yet, federated access to shell-based services comes with a number of challenges, since it typically requires local identities that need prior provisioning, as well as deprovisioning when no longer needed. Additionally, federated identities need to be securely mapped to local identities during authentication. Moreover, federated authorisation models based on Virtual Organisation (VO) membership, roles, and assurance levels need to be mapped as well to local privileges.

        Here we present our solution to these challenges, with a focus on OpenID Connect and SSH. This solution is applicable to other services as well, and is similarly being implemented for a webDAV service.

        In contrast to existing solutions for SSH that either require modified client and server software (GSI-based), or an additional step for obtaining additional credentials for the service (portal-based), the presented approach overcomes these limitations. Instead, we developed a set of client and server-side tools that complement but do not modify existing SSH software.

        The client tool (mccli), is implemented as a command line wrapper around the Unix-based SSH client. It enables on-the-fly account provisioning and transparent local account management: the users can directly log into the service with their federated credentials, without any prior application for an account. The server-side software consists of a lightweight web server (motley_cue) and a PAM module (pam-ssh-oidc). The key features of motley_cue are the mapping of federated to local identities with respecting site-local policies, as well as support for federated authorisation (VOs). The service administrators have full control over who is allowed to access their service, with the ability to only support certain VOs, filter users based on levels of assurance, and even specify authorised users individually. Via a generic and extensible interface, motley_cue is able to forward provisioning events into the local user management system (support exists for Unix accounts, Pooled Unix accounts, LDAP, and KIT user management). Admins can extend this to plug in their custom systems or username policies. Due to its modular design, motley_cue does not need to run on the same host as the service. SSH authentication uses PAM to prompt for an OpenID Connect Access Token and validate it via the REST API exposed by motley_cue.

        All software is free to use and is available on GitHub under MIT license, with support for the major Linux distributions. The software was tested with several major AAIs, such as EGI-Checkin, Helmholtz AAI, or DEEP IAM. Work is underway to enhance the service, including to add Windows support on the client side, deprovisioning, and extend support for other local user management systems.

        speaker info:
        https://www.linkedin.com/in/diana-gudu-aba07b10/?originalSubdomain=de

        Speaker: Diana Gudu (KIT)
      • 116
        Security considerations in the container-based environments.

        In the last years, the use of containers has expanded drastically. Many organisations are running containerised applications in cloud native environments. And since containers can encapsulate complex software
        stack with all the dependecies in an isolated enviroment (making the applications portable among clusters), they are widely used at many grid and HPC sites. How to know if these deployments are secure or how can a
        security expert or sysadmin, experienced in traditional server and VM based environments adapt to the container based deployments? In this talk EGI CSIRT will point out some of the most common security threats in the container-based environments together with the guidelines on how
        to prevent them.

        Speaker info:
        https://www.muni.cz/en/people/1388-daniel-kouril

        Speaker: Daniel Kouril (CESNET)
      • 5:10 PM
        Q&A
    • Workshop: Open Platforms for Digital Transformation of Earth Observation go.egi.eu/egi2021-2 (Zoom Room 2)

      go.egi.eu/egi2021-2

      Zoom Room 2

      The session presents the current state of the art of Digital Earth Observation by presenting the existing open platforms for processing Copernicus and other Earth Observation. The opportunities and limitations related to these Big Data solutions are discussed as background motivation of the federated C-SCALE approach that will help EOSC to play a role in the EO Digital Transformation. The architectural approach and service interfaces are illustrated through their use by example applications.

      Convener: Christian Briese (EODC)
    • 9:10 AM
      Chair Yoga go.egi.eu/egi2021-1 (Zoom Room 1)

      go.egi.eu/egi2021-1

      Zoom Room 1

    • 123
      Keynote: Climate modelling: data and computing challenges (Sylvie Joussaume, CNRS) go.egi.eu/egi2021-1 (Zoom Room 1)

      go.egi.eu/egi2021-1

      Zoom Room 1

      Sylvie Joussaume is an expert in climate modelling within CNRS. She has been involved in several assessment reports of the Intergovernmental Panel on Climate. Change. She coordinates since 2009 the European infrastructure project, IS-ENES, which integrates the European climate models in a common research infrastructure dealing with models, model data and high-performance computing for climate (http://is.enes.org), and which has integrated the EGI federation in 2020. She has been chair of the PRACE scientific committee in 2015. She also chairs the scientific evaluation committee of GENCI, the French national HPC facilities, and the scientific committee of ORAP that promotes high-performance computing in France.

      Speakers: Gergely Sipos (EGI.eu), Sylvie JOUSSAUME
    • 11:00 AM
      Break
    • Delivering Services and Solutions - Presentations go.egi.eu/egi2021-1 (Zoom Room 1)

      go.egi.eu/egi2021-1

      Zoom Room 1

      Convener: Silvio Pardi (INFN)
      • 124
        Full utilization of grid site computing resources using BOINC

        The CESNET grid site prague_cesnet_lcg2 supports several virtual organisations. These VOs use computing resources mostly intermittently and there were periods when CPUs were idle. We considered to add support to some LHC VOs, which submit jobs almost continuously. Since the size of resources is relatively small (1200 cores after the last upgrade) and the fraction for LHC VOs would be small when other VOs have running jobs, the overhead of fully supported WLCG site was high. LHC@Home project provides a convenient solution to this problem. ATLAS simulation jobs are run under CZ national BOINC account and the CPU contribution is accounted to the Czech Tier-2 site. We describe several setups tested and discuss experience with the final solution.

        Speaker info: https://www.fzu.cz/en/people/rndr-jiri-chudoba-phd

        Speaker: Dr Jiri Chudoba (CESNET)
      • 125
        Continuous integration and deployment on VIP

        The Virtual Imaging Platform (VIP, https://vip.creatis.insa-lyon.fr) is a web portal for medical simulation and image data analysis. It has been running for more than 10 years and has grown over the years to provide more and better services to researchers worldwide by leveraging the computing and storage resources of the EGI federation within the biomed Virtual Organization. On the internal and technical side, it has also become increasingly complex and this has made it harder for the developing team to maintain and to produce new features. Consequently we decided to enrich VIP with new continuous integration (CI) and continuous deployment (CD) practices in order to improve the software quality and to be more efficient in our computing tasks.

        The principle of CI/CD is to automate as much as possible to gain time and avoid human mistakes. For CI/CD on the VIP platform, we relied heavily on two dedicated tools : Jenkins (https://www.jenkins.io/) and Ansible (https://www.ansible.com/).

        Jenkins takes care of the CI part: it scans all our github code repositories and does several actions on every change. Essentially it verifies that the software code base compiles and it builds the final artifact that can be a binary executable, a library or a web service. It also launches the project test suite to validate the last developments and verify they do not add a bug or a regression. Test suites are very important in CI to ensure a very high software quality and to avoid as much as possible human validations that take a lot of time and are generally limited. At the moment we do not have enough tests on the VIP platform, but it will be an important focus in the future.

        Ansible takes care of the CD part, and allows to control perfectly how the different VIP software elements should be delivered and updated. The VIP architecture is complex, it has a web portal, a workflow engine with many plugins and it leverages resources from the Biomed Virtual Organisation of the EGI e-Infrastructure through Dirac. Before we set up Ansible, it was impossible to set up an environment independent and isolated from the production one to test new developments, making it difficult to test new features and to deliver new releases.
        Now, Ansible can easily and quickly deploy a complete VIP and ready-to-test platform on a temporary Cloud VM. Ansible also takes care of our new pre-production environment updating it automatically every time a new feature is validated.

        This presentation will explain why we set up new CI/CD tools and infrastructures in VIP, the important issues we had to deal with, and the benefits it brings. It is at the moment an internal and technical advancement making it easier and safer to add new features in VIP, but users will soon benefit from more frequent releases and a more reliable service.

        Speaker: Axel Bonnet (CNRS)
      • 126
        MSSEG-2: A medical imaging challenge on VIP

        Scientific challenges (competitions) bring together numerous research teams who work on solving a common scientific hard problem. During the challenge, their solutions are evaluated on a given set of data according to the guidelines given by the organizers.

        The automatic segmentation of tissus, structures and lesions in MRI brain scans is a challenging task for improving medical decision making. This is particularly prominent for the quantitative follow-up of Multiple Sclerosis (MS) patients. Then, the automatic detection of MS brain lesion was selected as a topic for a scientific challenge at the MICCAI conference in 2016. Following its success, and to evaluate the progress of computerized solutions accomplished in the last five years, especially impulsed by machine learning techniques, a second similar challenge (MSSEG-2, https://portal.fli-iam.irisa.fr/msseg-2/) is being organized at MICCAI 2021, this time for the detection of new MS lesions appearing between two patient's visits. The challenge uses data provided by OFSEP (http://www.ofsep.org/en) and the data management and processing infrastructure promoted by France Life Imaging (https://portal.fli-iam.irisa.fr/) comprising Shanoir and VIP.

        The Virtual Imaging Platform (VIP, https://vip.creatis.insa-lyon.fr) is a web portal for medical simulation and image data analysis. Hosted and provided by the Creatis laboratory, VIP is part of the FLI-IAM infrastructure and also a service of the EOSC (https://providers.eosc-portal.eu/service/creatis.virtual_imaging_platform) participating in the EGI-ACE project. VIP leverages resources available in the Biomed Virtual Organisation of the EGI e-Infrastructure to offer an open service to academic researchers worldwide. VIP currently counts more than 1300 registered users. Shanoir is used to store the anonymized patients images, the corresponding annotated images indicating the lesions (the ground truth hidden to the challengers) and the automatic differential lesion maps provided by the challengers algorithms processed for the final comparison.

        In the context of the MSSEG-2 challenge, the registered research teams provide their algorithms to the VIP team, who imports them into VIP and has them executed on the available resources. To facilitate the integration of new algorithms, we rely on containers and Boutiques descriptors, as described in the guidelines provided to the challengers : https://gitlab.inria.fr/amasson/lesion-segmentation-challenge-miccai21/-/blob/master/SUBMISSION_GUIDELINES.md

        At the time of submitting the abstract, algorithm integration is work in progress. 46 research teams registered to the challenge, i.e. possibly more than 46 algorithms to integrate in VIP. Although the VIP team has extensive experience with deploying applications on EGI resources, the organization of such a scientific challenge brings new challenges for us, too. The presentation will give more details on these challenges and the solutions deployed to successfully execute the algorithms provided by the challengers, such as access to GPU ressources, as well as the use of CVMFS and udocker.

        Speaker bio's:
        Sorina Pop is a CNRS research engineer at Creatis, currently in charge of the Virtual Imaging Platform (VIP). Since her Ph.D. degree in 2013, her activity has been focused on optimizing the execution of medical image processing applications on heterogeneous distributed systems. In the last few years, she has been particularly interested in enhancing open and reproducible science through her VIP activities, but also through other projects such as the EU OpenAIRE-Connect and EGI-ACE projects and the France Life Imaging (FLI) platform, where she is also member of the steering committee of the Information Analysis and Management node.

        Speakers: Dr Sorina Pop (CNRS), Axel Bonnet (CNRS)
      • 12:00 PM
        Q&A
    • Delivering Services and Solutions - Workshop: EGI-ACE Federated data access services go.egi.eu/egi2021-4 (Zoom Room 4)

      go.egi.eu/egi2021-4

      Zoom Room 4

      The session will describe the different services available in EGI-ACE for Federated Data access, including the detailed roadmap for each service and achievements.
      The session will be chaired by Andrea Manzi (EGI Foundation), with contributions from the different service responsibles from UKRI-STFC, Cyfronet, Enhance-R.
      The session will include an example of data space in EGI-ACE (Prominence) which is integrating the Federated data access services in EGI-ACE

      Convener: Andrea Manzi (EGI.eu)
    • Innovating Services Together - Workshop: HPC integration in EGI-ACE go.egi.eu/egi2021-3 (Zoom Room 3)

      go.egi.eu/egi2021-3

      Zoom Room 3

      EGI-ACE aims at providing interoperability guidelines for HPC systems with the EOSC Cloud Compute platform. Four scientific pilot use cases with combined cloud and HPC needs will be used for exploring and identifying how HPC systems should be exposed to the EOSC portal and how users should interact with them.

      This session will provide a status update with presentations from the scientific pilots, HPC centres and technology providers involved in the activity

      Convener: Enol Fernandez (EGI.eu)
      • 134
        HPC integration in EGI-ACE

        EGI-ACE aims at providing interoperability guidelines for HPC systems with the EOSC Cloud Compute platform. Four scientific pilot use cases with combined cloud and HPC needs will be used for exploring and identifying how HPC systems should be exposed to the EOSC portal and how users should interact with them.

        This session will provide a status update with presentations from the scientific pilots, HPC centres and technology providers involved in the activity

        Speaker: Enol Fernandez (EGI.eu)
      • 135
        ssh-oidc at TUBITAK

        Hakan Bayindir (Male) is working as a senior researcher at TUBITAK ULAKBIM since 2006. He has worked in many European Projects such as EGI, EGI-InSpire, EOSC-hub, and EGI-ACE. He has many hats ranging from system administration to managing projects themselves and has experience in cluster and cloud system design & administration, high performance programming, development of simulation systems, and project management. He completed both FitSM Advanced Certifications, and currently oversees EGI-ACE project for TUBITAK ULAKBIM.

        Speaker: Hakan Bayindir (TUBITAK)
      • 136
        Running containers in advanced computing environments with udocker

        udocker is a tool to enable execution of Linux containers in user space. Distinctively from other tools, udocker is meant for easy deployment, and provides multiple execution engines to cope with different host environments. udocker does not require privileges and can execute containers with or without Linux namespaces making it suitable to execute containers across a wide range of Linux computing environments such as HPC resources.

        Speaker: Jorge Gomes (LIP)
      • 137
        Experiences in integration with HPC @ INFN

        The talk will provide information about the activities going on at INFN to leverage some HPC compute clusters: Cineca, CNAF and ReCaS-Bari. There will be few details about which approach, solutions and technologies are used currently to include also HPC resources in the Compute Infrastructure used at INFN

        Speaker: Tommaso Boccali (INFN)
      • 138
        HPC benchmarking pilot

        This presentation will show progress on benchmarking, containerization, and investigations on data access performed over the past months to enable heterogeneous support for HEP workloads at HPC sites.

        Speaker: David Southwick
      • 139
        A HPC as a Service solution for climate data analytics on the EGI infrastructure

        The deluge of data started at the beginning of this century has caused a profound transformation in the way scientific discovery is carried out. In several domains, such as climate science, scientific advances now rely on technologies and software solutions from both the HPC and Big Data landscapes. However, being able to efficiently exploit HPC infrastructures for running scientific data analysis is not easy. A unified model that also allows the deployment on HPC of the same services already exploited in the cloud can pave the way for a wider range of opportunities in the scientific community, further fueling the adoption of the HPC as a Service (HPCaaS) paradigm.

        In this respect, software containers are good candidates for supporting portability and deployment of data analytics frameworks over multiple platforms. Thanks to the recent development of HPC-friendly container technologies (e.g., udocker, Singularity, Sarus), it has now become possible for scientists to exploit the benefits of this model also on HPC infrastructures.

        Containers can allow encapsulation of the application, together with its dependencies, into a single and portable image file. Nevertheless, several issues must be addressed to effectively exploit these technologies for operational scientific services and applications. Specifically, some of the main issues concern possible performance degradation and the integration with services, including those distributed across different infrastructures.

        In the European Open Science Cloud (EOSC) context, the ENES Climate Analytics Service (ECAS) is a central component of the ENES Data Space set up in the EGI-ACE project, which aims to provide an open and cloud-enabled data science environment for climate scientists. In this environment, the Ophidia HPDA framework represents the core computing engine of the ECAS service and it can greatly benefit from the exploitation of HPC resources for running data analysis applications and workflows.

        This work presents the container-based approach investigated in the EGI-ACE project, for transparent and portable deployment of ECAS on top of the HPC resources made available in the EGI infrastructure.

        "Fabrizio Antonio was awarded a Master’s Degree with first-class honors in Computer Engineering from the University of Salento, Faculty of Engineering, in April 2016, with a thesis on High Performance Computing and Big Data. In May 2016, he joined the Data Science and Learning Research Team within the Advanced Scientific Computing (ASC) Division of CMCC.
        His research activities focus on distributed data management and high-performance data analytics and mining for eScience in the context of climate change."

        Speaker: Mr Fabrizio Antonio (CMCC Foundation, Lecce, Italy)
    • Innovating Services Together - Workshop: Service Portfolio evolution go.egi.eu/egi2021-2 (Zoom Room 2)

      go.egi.eu/egi2021-2

      Zoom Room 2

      This session will outline the evolution of the EGI service portfolio, covering the internal federation services as well as the external services and those drawn in from individual federation members.

      It will include updates from the inclusion of services from the EGI ACE project and work on a Service Strategy and Service Roadmap for EGI.

      The session will have two presentations, one on the EGI foundation delivered services and one on the larger federation services, before breakouts on directions for service evolution.

      Convener: Owen Appleton (EGI.eu)
      • 140
        The strategic landscape for EGI services

        Considering the landscape in which EGI services sit, and how it is evolving through EOSC and other intitiatives.

        Speakers: Owen Appleton (EGI.eu), Sergio Andreozzi (EGI.eu)
        • a) The strategic landscape for EGI services

          How EGI services fit into the larger European landscape

          Speaker: Owen Appleton (EGI.eu)
      • 141
        Community Services from EGI-ACE and Beyond

        Introducing the wider services from the EGI federation being drawn into the broader EGI Portfolio

        Speaker: Gergely Sipos (EGI.eu)
      • 142
        Introduction to the Strategic Objectives and breakouts

        Introducing the new Service Strategy draft and the breakouts

        Speaker: Sergio Andreozzi (EGI.eu)
      • 143
        Breakout sessions

        There will be 4 breakout sessions to discuss 4 draft strategic objectives for the EGI Service Strategy.

        Breakout session 1:
        Facilitator: Gergely Sipos
        SO1: Federate and deliver a compute continuum from edge, through cloud and cluster to HPC together

        Breakout session 2:
        Facilitator: Mark Dietrich
        SO2: Implement and deliver research ‘Data Spaces’ as a joint effort with Research Infrastructures

        Breaksession 3:
        Facilitator: Ville Tenhunen
        SO3: Make the EGI platform AI/ML-ready and lower barriers to entry for researcher using AI/ML

        Breakout session 4:
        Facilitator: Owen Appleton
        SO4: Deliver a federated trusted computing platform for sensitive data

        After a short introduction to the SO, participants will have the opportunity to provide feedback in a collaborative way using Mural.

        Speakers: Gergely Sipos (EGI.eu), Mark Dietrich (EGI.eu), Owen Appleton (EGI.eu), Sergio Andreozzi (EGI.eu), Ville Tenhunen
      • 144
        Open Discussion: changes and additions to strategic objectives
        Speaker: Owen Appleton (EGI.eu)
    • 12:45 PM
      Break
    • 12:45 PM
      Lunch
    • Demonstration: Notebook-as-a-VRE (NaaVRE): Scaling Jupyter Notebooks go.egi.eu/egi2021-4 (Zoom Room 4)

      go.egi.eu/egi2021-4

      Zoom Room 4

      • 145
        Notebook-as-a-VRE (NaaVRE): Scaling Jupyter Notebooks

        The study of many scientific problems concerning environmental challenges sets significant computational demands, such as large data volumes, advanced modeling techniques, and distributed computing facilitates. To conduct such investigations, a researcher often has to reuse virtual assets, e.g., observational data or images, AI models, operational workflows, and infrastructure services from different parties, for building computational experiments. Jupyter in such a scenario, allows researchers to effectively implement their experimental logic using scripting languages to document and share experiments with their necessary inputs and parameters.
        In the case of ecological and biodiversity scientists, AI and statistical models are used to analyze a host of different observations, such as specimen records, citizen science observations, eDNA, and remote sensing imaging. For example, high-resolution Light Detection and Ranging (LiDAR) datasets are widely used to monitor changes in an ecosystem structure, and to predict that the distribution of species across space and time. Most often, Jupyter notebooks are used in this analysis,
        as they allow researchers to effectively implement an experimental logic using languages such as Python. However, Jupyter faces challenges of utilizing remote infrastructure:
        Difficulty to find and reuse a notebook at the cell level. The lack of metadata hampers the discovery of useful fragments of code (namely Cells).
        Lackingflexibility and portability to reuse a notebook as part of a workflow. The tightly coupled functions of a notebook, i.e. library dependencies make the reusability and portability of the code fragments difficult.
        Difficulty to scale the notebook to remote infrastructures. Current notebook environments, like Jupyter Hub, use pre-configured infrastructures. When processing huge data volumes or computationally complex tasks, dynamically allocated cloud resources are needed for parallelizing distributed computing tasks.
        To tackle those challenges, we propose a VRE solution that can be embedded into Jupyter as an extension that enables exporting individual cells as docker containers that can be composed as workflows. Our solution is composed of the following components:
        Containerizer: It tracks the user’s interactions with the notebook and in real-time updates the code and metadata based on the cell’s modified content. When a user wishes to publish their cell, the containerizer builds a docker image off-premise with its metadata using our infrastructure automator called Software Defined Infrastructure Automator (SDIA).
        Experiment manager: This component is responsible for loading and editing the cells’ metadata catalog as well as for the submission and monitoring of workflows on the provisioned infrastructure.
        Software-Defined Infrastructure Automator (SDIA). SDIA automates the planning, provisioning, monitoring, and adaptation of applications and their infrastructure on multi-cloud provider cloud offerings.

        About the speakers:

        Spiros Koulouzis is a researcher at the University of Amsterdam. His research interests include scientific workflows, as well as distributed and parallel systems. Spiros has a Ph.D. in computer science from the University of Amsterdam.

        Zhiming Zhao is currently a senior researcher in the group of System and Network Engineering ( SNE) at University of Amsterdam ( UvA). He obtained his bachelor and master degrees in Computer Science from Nanjing Normal University ( NJNU) and East China Normal University ( ECNU) in 1993 and 1996 in China respectively. He obtained his Ph.D. in Computer Science from University of Amsterdam (UvA) in 2004. He has strong research interest in advanced computing and network technologies, time critical and data intensive systems, Cloud computing, scientific workflows and software agents. He coordinates research and development activities in the EU H2020 project SWITCH (Software Workbench for interactive time critical and highly self-adaptive cloud applications), and in the “Data for Science” theme in the EU H2020 environmental science cluster project ENVRIPlus. He also leads the research tasks of research sustainability in the EU H2020 VRE4EIC project, and of semantic linking in the EU FP7 ENVRI project.

        Speakers: Spiros Koulouzis (University of Amsterdam), Zhiming Zhao (EGI.eu)
    • Demonstration: Serverless computing across the Cloud continuum for Deep Learning Inference with OSCAR go.egi.eu/egi2021-4 (Zoom Room 4)

      go.egi.eu/egi2021-4

      Zoom Room 4

      • 146
        Serverless computing across the Cloud continuum for Deep Learning Inference with OSCAR

        OSCAR is an open-source platform to support serverless computing for compute-intensive data-processing applications. OSCAR runs on dynamically provisioned auto-scaled Kubernetes clusters deployed through the Infrastructure Manager (IM), an open-source Infrastructure as Code (IaC) tool. These Kubernetes clusters include support for MinIO, an open-source object storage, which fires events in response to file uploads in order to trigger the file processing as Kubernetes jobs. The clusters can grow and shrink thanks to CLUES, an open-source elasticity manager that deploys additional nodes through the IM and terminates them whenever they are no longer needed.

        OSCAR has evolved in the last years to tighten the integration with SCAR, an open-source tool to execute containers in AWS Lambda with automatic delegation into AWS Batch. The use of a Functions Definition Language (FDL) allows to define data-processing serverless workflows that can perform some processing in an on-premises Cloud while delegating the most computationally-intensive part in a public Cloud such as Amazon Web Services (AWS).

        OSCAR is integrated with the IM Dashboard to facilitate the deployment of OSCAR clusters across a myriad of Cloud providers including major public clouds, widely used Cloud Management Platforms and federated infrastructure such as the EGI Federated Cloud. Thanks to the integration with EGI Check-In, users can seamlessly access the IM dashboard to self-provision these clusters.

        In order to expand the variety of use cases supported, two additional features have been recently implemented. First, synchronous support with scale-to-zero capabilities. This allows synchronous data-processing functions packaged as Docker containers that can be invoked through both a REST API and the CLI. Second, OSCAR can run on minified Kubernetes distributions (such as K3s) in order to execute on low-powered devices such as Raspberry PIs. This is required for use cases that require executions in the edge, for lightweight processing, such as inference of previously trained Deep Learning (DL) models.

        Together, the integration of OSCAR / SCAR provides an open-source platform that supports serverless computing across the Cloud continuum, where execution can take place in low-powered devices, in on-premises Clouds and in federated or public Clouds. In this contribution we will demonstrate how a user can seamlessly provision an OSCAR cluster using the IM Dashboard in the EGI Federated Cloud in order to create a serverless workflow for mask detection from a trained DL model in public crowds across different infrastructures and using EGI DataHub as one of the storage back-ends.

        Speaker bio:

        Sebastián Risco received a BSc degree in Computer Engineering from the Universitat Politècnica de València (UPV), Spain, in 2017. In 2017 he started his MSc degree in Information Management. He joined the Grid and High Performance Computing research group (GRyCAP) in 2018, while he worked on his Master’s Thesis. His research interests are focused in Serverless Computing, Cloud Computing and Container Orchestration Systems.

        Speaker: Sebastián Risco (UPVLC)
    • Delivering Services and Solutions - Workshop: EGI Workload Manager Service go.egi.eu/egi2021-3 (Zoom Room 3)

      go.egi.eu/egi2021-3

      Zoom Room 3

      The EGI Workload Manager service provides an easy access to the EGI computing and storage resources for users from various scientific domains. It helps to increase the efficiency of grid and cloud computations for individual users with moderate requirements as well as for large user communities needing a powerful workflow management system.
      The Service is built with the software provided by the DIRAC Interware project. The software is rapidly evolving in order to accommodate new technologies in grid and cloud computing including new types of resources, new security frameworks, etc. Users can access the service with a variety of interfaces, including the Web interface. The DIRAC API’s allow creation of specific application portals taking a full advantage of the workflows management by the EGI Workload Manager.
      During the proposed session we will provide an overview of the service main features and capabilities, user interfaces and the overall status of the project evolution. Recent developments in the EGI cloud integration, OAuth/OIDC AAI frameworks and others will be discussed. Examples of the service practical usage by several user communities will be presented. In particular, we envisage reports from the WeNMR and Biomed communities which have a long-lasting experience with the use of the EGI Workload Manager.
      The goal of the session is to attract new users by showing advantages offered by the service and to demonstrate how to get started quickly with its usage. We will discuss typical problems of new user communities and will propose appropriate solutions

      Convener: Andrei Tsaregorodtsev (CNRS)
      • 147
        DIRAC Interware Project Status

        The DIRAC Interware Project provides a software framework and develops ready to use components to build distributed computing systems of arbitrary complexity. The solution is universal and covers both workload and data management tasks. The Project was originally developed for LHCb – one of the 4 experiments at the LHC collider which stays its principal user. Therefore, a special care is taken to ensure massive operations with user computational tasks and data.
        In this contribution we will give an overview of the Project, its main components and interfaces. We will present recent updates of the DIRAC software such as migration to Python3, new service APIs, software deployment, support for new authentication and authorization frameworks and other topics. We will focus on practical aspects to help users in finding the most efficient way interaction with the DIRAC services provided by the EGI infrastructure.

        Speaker: Andrei Tsaregorodtsev (CNRS)
      • 148
        EGI Workload Manager service

        The EGI Grid Infrastructure Workload Manager Service (EGI-WMS) is based on the DIRAC Interware distributed computing framework. It allows, through pilot jobs, to provide diverse scientific communities with a generic access to heterogeneous computing and storage resources.
        This presentation will describe the service platform and the hosting site migration that was performed at the beginning of the EGI-ACE project (during 2021-Q1) and the resulting changes for the user communities. We will then briefly explain how a new user may start submitting jobs on the WMS, and finally some use cases will be described giving some examples of the different computing models managed and specific workflow needs satisfied by the service.

        Speaker: Gino Marchetti (CNRS / CC-IN2P3)
      • 149
        WeNMR project: boosting structural biology research with DIRAC services

        The WeNMR Collaboration is a virtual research community which brings together several research teams in structural biology providing a common platform for data analysis and modelling. WeNMR is one of the most active users of the EGI Workload Manager service. User tasks prepared by the WeNMR application portals are submitted to the Workload Manager service which then dispatches them to the currently available grid sites. The pilot job based workload scheduling architecture allows to easily add new resources transparently for the users. It also makes it easy to apply usage policies by defining fine grained priorities to certain activities.
        In this contribution we will describe how the WeNMR application portals are interfaced to the DIRAC services and provide examples of usage including allocation and accounting of resources for the COVID’19 related studies.

        Speaker: Alexandre Bonvin (eNMR/WeNMR (via Dutch NGI))
      • 150
        Biomedical community accessing grid resources via the EGI Workload Manager

        The biomed Virtual Organization (VO) is a large scale international and multi-disciplinary VO supporting communities from the Life Sciences sector, with three main thematic groups: medical image analysis, bioinformatics and drug discovery. The VO is operated on the EGI infrastructure and supported by more than 50 sites, delivering access to a large number of heterogeneous resources.
        The Virtual Imaging Platform (VIP, https://vip.creatis.insa-lyon.fr/) is a web portal enabling biomed researchers worldwide to access multiple applications, as well as important amounts of storage and computing simply through their web browser. VIP relies on Dirac for both job submission and data management on EGI biomed resources. VIP currently counts more than 1300 registered users, some 20 applications and 61 publications with results computed with VIP since 2011.
        Based on our experience, we will present the Virtual Imaging Platform, one of its latest success stories and the way it is interfaced with Dirac and EGI.

        Speakers: Axel Bonnet (CNRS), Sorina POP (CNRS)
      • 151
        Demonstration, Question and Answers

        In this contribution we will make a demonstration of typical user tasks when working with the EGI Workload Manager service. This is an interactive session when all the necessary explanations will be given and user’s questions will be answered.

        Speaker: Andrei Tsaregorodtsev (CNRS)
    • Envisioning the Future: Expanding EGI Business Models go.egi.eu/egi2021-1 (Zoom Room 1)

      go.egi.eu/egi2021-1

      Zoom Room 1

      This session will introduce the EGI Business Model Innovation Program as a mechanism for defining different business models to deliver value, which are validated through concrete business cases.

      The objective of the program is to capture opportunities for the EGI Federation where financial transactions will be foreseen and not yet captured as part of the established mechanisms in order to ensure that all strategic, legal, and administration aspects have been analysed.

      During this session a number of ongoing concrete use cases will be presented (BMI projects) as well as an opportunity for discussion, feedback and contributions.

      Convener: Sy Holsinger (EGI.eu)
    • Innovating Services Together - Workshop: EGI Technical Architecture and roadmap go.egi.eu/egi2021-2 (Zoom Room 2)

      go.egi.eu/egi2021-2

      Zoom Room 2

      This session focuses on the presentation of the EGI Technical Annual Plan that includes all the planned activities related to the innovation of existing EGI services and the piloting of new services and solutions. The EGI Technical Annual Plan is developed by the Technical Solution Team in collaboration with technology providers and research & innovation projects for technology scouting and expert user groups for requirements gathering.
      Innovation on the EGI services will be presented per technical area (Compute, Data Management, AAI, emerging technical areas) with relevant examples of use cases that will benefit of the evolution of the EGI service offer. Furthermore, current pilot activities to assess new services and technologies will be introduced.

      Convener: Diego Scardaci (EGI.eu)
    • 162
      Keynote: 25 years of CESNET – history and future challenges (Helmut Sverenyák, CESNET) go.egi.eu/egi2021-1 (Zoom Room 1)

      go.egi.eu/egi2021-1

      Zoom Room 1

      In this presentation Helmut will present the development and achievements of the CESNET from NREN to national e-infrastructure provider through the past 25 years.

      Helmut Sverenyák, Research and Development Manager – has a Dipl. Ing. (1991) from the Faculty of Nuclear Sciences and Physical Engineering at the Czech Technical University in Prague. He joined CESNET in 1998 as a member of the Research and Development department. In 2004, he was made head of Research and Development, responsible for the coordination and support of R&D activities and projects managed by CESNET. In the period 2012–2014 Helmut was a member of TERENA Executive Committee. Since 2015 he has been the Vice Chair of the ELIXIR CZ (Czech national ELIXIR infrastructure node) board.

      Speaker: Helmut Sverenyák (CESNET)
    • 163
      EGI Federation looking forward: a summary of the Conference outcomes go.egi.eu/egi2021-1 (Zoom Room 1)

      go.egi.eu/egi2021-1

      Zoom Room 1

      The EGI Conference gives the voice to research communities, technology and infrastructure providers to steer the evolution of scientific computing and the EGI Federation in Europe.
      Gergely Sipos, Head of Services Solutions and Support at the EGI Foundation, will summarize the conference outcomes with the help of rapporteurs.

      Speaker: Gergely Sipos (EGI.eu)
    • 164
      Closing ceremony go.egi.eu/egi2021-1 (Zoom Room 1)

      go.egi.eu/egi2021-1

      Zoom Room 1

      Speaker: Tiziana Ferrari (EGI.eu)
    • GOSC: The Second Global Open Science Cloud (GOSC) workshop go.egi.eu/egi2021gosc

      go.egi.eu/egi2021gosc

      Zoom

      While the celebration of the science and technology at the EGI Conference is closing down, carrying on with the enthusiasm, we warmly welcome you to the 2nd GOSC Workshop!

      Together with EGI Foundation, the Computer Network Information Center (CNIC) of the Chinese Academy of Science, and CODATA, we welcome you to join the Second Global Open Science Cloud (GOSC) workshop, aiming to give an overview of GOSC activities and discuss the potential collaboration through GOSC. We invited the well-developed CODATA GOSC Working Groups and Case Studies Groups. It will be the first time for them getting together to share group discussions and activities. We also invite global advanced digital infrastructures practices Open Science including, OSG in US, Australian Research Data Common (ARDC), Chinese Science and Technology Cloud (CSTCloud), Malaysia Open Science Platform (MOSP), African Open Science Platform (AOSP), and South African Open Science Cloud (SAOSC). This will be an extension of European Open Science Cloud (EOSC) discussion during the conference, and will be a unique opportunity to review the global open science infrastructure development, share experiences, and identify concrete collaborations.

      The workshop will have 2 x 90min sessions:
      1. GOSC activities and development
      2. Global collaborations through GOSC

      We welcome all to join, together we can build this new GOSC community. In particular, we invite attendance by:

      • Research community and research infrastructure representatives with needs and experience supporting global collaborations;
      • Digital infrastructure representatives open to participating in a global resource federation;
      • CODATA GOSC Working Group Members
      • Experts on e-infrastructure and technology developing and operating solutions for federated access to data, computing, software and applications.

      This a follow-up of the success of the 1st GOSC workshop in Nov 2020. A report outcome from the 1st GOSC workshop is now published which reviews the GOSC landscape: https://zenodo.org/record/5575275#.YW198hBBwUR

      We are honoured to invite our favoured speakers as below:
      Dr Simon Hodson has been Executive Director of CODATA since August 2013. He has contributed to influential reports on Current Best Practice for Research Data Management Policies and to the Science International Accord on Open Data in a Big Data World. He chaired the European Commission’s Expert Group on FAIR Data which produced the report Turning FAIR into Reality. He is currently vice-chair of the UNESCO Open Science Advisory Committee, tasked with drafting the UNESCO Recommendation on Open Science, which is intended for adoption in November 2021. As a significant part of his CODATA role, Simon is tasked with preparing a major ISC and CODATA Decadal Programme on ‘Making Data Work for Cross-Domain Grand Challenges’, which will improve the coordination of specifications for data integration and interoperability for interdisciplinary research.

      Prof Jianhui Li is the director of Science and Technology Cloud Department at the Computer Network information Center (CNIC) of the Chinese Academy of Sciences (CAS), and a Professor at the University of Chinese Academy of Sciences (UCAS). He obtained PhD on computer science from the Institute of Computing Technology of CAS in 2007. In 2016, He founded “China Scientific data”, which is the first open access data journal for scientific data publication in China. Currently, he is leading the design, development and operation of CSTCloud (China Science and Technology Cloud), which is the national level open science platform. He also serves as the CODATA vice president.

      Prof Miron Livny is a senior researcher and professor specializing in distributed computing at the University of Wisconsin–Madison. Livny has been a professor of computer science at Wisconsin since 1983, where he leads the HTCondor high-throughput computing system project. Miron is also a principal investigator and currently the facility coordinator for the Open Science Grid project, Director of the Center for High Throughput computing, CTO of Wisconsin Institutes for Discovery, and Director of Core Computational Technology of the Morgridge Institute for Research. In 2006, along with Raghu Ramakrishnan, Professor Livny won the SIGMOD Test of Time award for his seminal work on distributed databases.

      Dr Carmel Walsh is the Director, eResearch Infrastructure & Services at the Australian Research Data Commons (ARDC). The ARDC provides Australian researchers with competitive advantage through data. It does this by enabling the Australian research community and industry access to nationally significant, data intensive digital research infrastructure, platforms, skills and collections of high quality data. Carmel leads on storage and compute with a focus on the national research cloud compute service for Australian researchers, the ARDC Nectar Research Cloud, and national Data Retention project.

      Dr Happy Sithole is the Centre Manager for the South African cyberinfrastructure (CI) system, NICIS since 2019. He is responsible for the coordination of this system so that it meets the demands of the South African National System of Innovation, in support of the Department of Science and Innovation. This involves the three main entities of the CI, Centre for High-Performance Computing (CHPC), South African National Research Network (SANReN), Data Intensive Initiative of South Africa (DIRISA). Prior to becoming the CI centre manager, Sithole spent 12 years as the director of the Centre for High Performance Computing. He completed his PhD in materials science at the University of Limpopo.

      Dr Ingemar Häggström is the Head of operations, EISCAT Scientific Association. EISCAT is running the only incoherent scatter radars in Europe for atmospheric and geospace research. A new system, the imaging radar EISCAT_3D, and are now being built. He presents EISCAT and EISCAT_3D in the global scale of geospace research, and how scientists can interact with and access the resources of the infrastructure. This involves interchanges with similar radar facilities around the world and a system for setting up a common (meta)data federation, federated processing and data movements.

      Dr Anwar Vahed is the director of the Data Intensive Research Initiative of South Africa (DIRISA). He manages the South African national research data infrastructure supporting data intensive research, and coordinates the development of relevant expertise and strategies, as well as the implementation of policies for research data management. DIRISA, together with the Centre for High Performance Computing (CHPC) and the South African Research Network (SANReN), comprises the technology area of the National Integrated Cyberinfrastructure System of South Africa. is the director of the Data Intensive Research Initiative of South Africa (DIRISA). He manages the South African national research data infrastructure supporting data intensive research, and coordinates the development of relevant expertise and strategies, as well as the implementation of policies for research data management. DIRISA, together with the Centre for High Performance Computing (CHPC) and the South African Research Network (SANReN), comprises the technology area of the National Integrated Cyberinfrastructure System of South Africa.

      Dr Milan Ojsteršek is the Head of Laboratory for heterogenous computer systems, FERI, University of Maribor. He received the PhD in computer science from the University of Maribor in 1994. Milan leads the Laboratory for heterogenous systems at the Faculty of Electrical Engineering and Computer Science on University of Maribor. His research and project work focuses on heterogeneous computing systems, digital libraries, natural language processing, web application development, web technologies, knowledge management, semantic web and service-oriented architecture.

      Dr Lili Zhang is a research scientist at the Computer Network Information Center of Chinese Academy of Sciences and also a member of CODATA International Data Policy Committee. Her research focuses on open data and open science policy, practice; information economics. Lili is the co-chair of CODATA GOSC Policy Working Group, SDG-13 Case Study Working Group. She will present the work plan and activities happened in these working groups.

      Dr CheeSun Liew is a senior Lecturer at the Department of Computer System & Technology, Faculty of Computer Science & Information Technology, University of Malaya. He received a PhD in Computer Science from Edinburgh University. His research focuses on data-intensive computing, Grid, High Performance Computing etc. CheeSun is involved in Malaysia Open Science Platform (MOSP) and will introduce the development of Open Science in Malaysia.

      More information

      Convener: Yin Chen (EGI.eu)