EGI2023

Europe/Amsterdam
Novotel Poznań Centrum

Novotel Poznań Centrum

pl. Andersa 1 61-894 Poznań Poland
Description

EGI2023

Where Research Meets Computing

EGI2023 took place in Poznań, Poland - from June 19th to June 23rd. 

At the annual EGI conference, international scientific communities, computing and service providers, European projects, security experts, community managers, and policy makers gather to take research and innovation in data-intensive processing and analytics forward.

Looking for more pictures? 

 

 

Contact
    • 09:00 17:00
      CSIRT: Closed Meeting
    • 09:00 17:00
      EOSC DIH Bootcamp: Training

      Pitch to win
      Trainer: David Beckett (Best 3 Minutes)

      This practical, interactive pitch training session is designed to help members of the companies, startups, spin-off and in general members of the EOSC DIH community improve their ability to deliver concise and effective pitches.

      Participants will learn how to create a compelling pitch that effectively communicates their message, captures the attention of their audience, and motivates action. Through interactive exercises participants will learn how to;

      Tune their pitch to the Audience, and be clear on the Objective
      Create a pitch storyline using The Pitch Canvas© and brainstorming techniques
      Have a strong Opening and Closing of the pitch
      Use the Power of Three to keep focused and memorable
      Learn how to practice and raise the level in terms of storyline and delivery
      Objectives:

      The objective of this pitch training session is to offer companies, SMEs, startups, spin-offs and other members of the EOSC DIH community some guidelines and hands-on training to:

      Refine the Pitch, developing a clear and concise message that effectively communicates the value proposition, progress, team and ask, with a clear Call to Action
      Improving Communication Skills, developing effective presentation skills
      Enhancing Investor Appeal, identifying the key factors that investors are looking for in a startup and tailoring the pitch to meet those expectations.
      Increasing Confidence, helping to overcome any nervousness or anxiety that they may feel when presenting their ideas to potential investors or customers.
      Manage the limited time available for your pitch!
      Afterwards, you'll get access to a selection of support materials to help you put what you learned in the workshop into practice over the following weeks and months. 

      Training on Using the EOSC Portal for business and Workshop on User Experience

      Trainer: Adam Sipowicz (Abroco)
      This workshop on user experience of the private sector as users of the European Open Science Cloud (EOSC) aims to bring together stakeholders from academia, industry, and the public sector to discuss and exchange ideas on how to improve the user experience of private sector users on the EOSC platform.

      Through interactive sessions, participants will share their experiences and challenges in using the EOSC platform, as well as explore best practices and strategies for enhancing user experience. The workshop will also provide insights into the unique needs and requirements of private sector users, and how they can be better integrated into the EOSC ecosystem.

      This bootcamp is a local event of Data Week 2023.

      • 09:00
        EOSC DIH Bootcamp: Pitch To Win 3h

        Pitch to win
        Trainer: David Beckett (Best 3 Minutes)

        This practical, interactive pitch training session is designed to help members of the companies, startups, spin-off and in general members of the EOSC DIH community improve their ability to deliver concise and effective pitches.

        Participants will learn how to create a compelling pitch that effectively communicates their message, captures the attention of their audience, and motivates action. Through interactive exercises participants will learn how to;

        Tune their pitch to the Audience, and be clear on the Objective
        Create a pitch storyline using The Pitch Canvas© and brainstorming techniques
        Have a strong Opening and Closing of the pitch
        Use the Power of Three to keep focused and memorable
        Learn how to practice and raise the level in terms of storyline and delivery
        Objectives:

        The objective of this pitch training session is to offer companies, SMEs, startups, spin-offs and other members of the EOSC DIH community some guidelines and hands-on training to:

        Refine the Pitch, developing a clear and concise message that effectively communicates the value proposition, progress, team and ask, with a clear Call to Action
        Improving Communication Skills, developing effective presentation skills
        Enhancing Investor Appeal, identifying the key factors that investors are looking for in a startup and tailoring the pitch to meet those expectations.
        Increasing Confidence, helping to overcome any nervousness or anxiety that they may feel when presenting their ideas to potential investors or customers.
        Manage the limited time available for your pitch!
        Afterwards, you'll get access to a selection of support materials to help you put what you learned in the workshop into practice over the following weeks and months.

      • 14:00
        Training on Using the EOSC Portal for business and Workshop on User Experience 3h

        Trainer: Adam Sipowicz (Abroco)
        This workshop on user experience of the private sector as users of the European Open Science Cloud (EOSC) aims to bring together stakeholders from academia, industry, and the public sector to discuss and exchange ideas on how to improve the user experience of private sector users on the EOSC platform.

        Through interactive sessions, participants will share their experiences and challenges in using the EOSC platform, as well as explore best practices and strategies for enhancing user experience. The workshop will also provide insights into the unique needs and requirements of private sector users, and how they can be better integrated into the EOSC ecosystem.

        This bootcamp is a local event of Data Week 2023.

    • 09:00 17:00
      FitSM Training: Training
    • 14:00 15:30
      Building Stronger Communities with Next-Generation Access and Identity Management: Training

      In this training event, you will acquire knowledge of the fundamentals of Identity and Access Management (IAM) in the EGI ecosystem.  IAM is a very important part of understanding how to securely and efficiently manage the services and users with an EGI Check-in account. This course will explore the following components of IAM with some hands-on exercises:

      What is identity and access management
      Concept of community, identity and token
      Relationship between IDP provider, service provider, relying party
      What are Open ID/SAML/OAuth protocols and related flows
      Access control and review of user access
      Provisioning and Deprovisioning
      Who this course is for: 

      Whether or not you are a member of a virtual community or you are a community manager, this course is designed for you!

    • 14:00 15:30
      Installation and configuration of ARC-CE on HPC: Training

      The tutorial demonstrates the installation and configuration of an ARC-CE front-end for use in a distributed grid infrastructure, such as WLCG. Particular focus will be on supporting high-performance systems, using experience from Vega EuroHPC, Nordic WLCG Tier1 and other HPC centres. The tutorial addresses primarily system administrators, but also serves as a demonstrator of a seamless access to HPC resources to extended EOSC user communities, such as Life Sciences, Climate and Biodiversity, Astrophysics, Materials science and others.

      The tutorial will demonstrate the installation of the not yet released ARC 7, and focusing on an ARC-CE set up to support tokens.

      The tutorial material can be found on the ARC documentation pages. 

      A handful of test-clusters have been set up if attendees would like to follow along. Sign up for a cluster by adding your name in the following google spreadsheet:https://docs.google.com/spreadsheets/d/1C5E9LCIOayxcoQQlRrqXEBnFhAuZu6Ov9Mexf1MmB-A/edit?usp=sharing

      We are looking forward to seeing you!

      Check the presentation of the ARC technical overview.
      Check the presentation on the Installation and configuration of ARC-CE on VEGA

    • 14:00 15:30
      Open Science (by OpenAIRE) 1h 30m

      To address questions about researchers’ obligations as a Horizon Europe (HE) grant holder in terms of Open Access to publications and Research Data Management, this workshop will outline those policies. We will present these HE requirements, followed by some time for Q&A. Participants will also be introduced to the main tools and services, including those from OpenAIRE, that provide help to project coordinators and research support staff on the requirements' compliance.

      In addition, this workshop will guide participants on HE data management plans (DMPs), and how to write these through practical examples, following their requirement set by the EC, and which will be made achievable by a better understanding of HE Open Science policies.

      Learn about:

      Mandatory and recommended Open Science requirements in HE.
      Compliance with the HE Open Access to publications mandate.
      Managing and sharing Research Data in HE projects.
      Delivering DMPs and reporting publications and datasets in HE.
      Tools and services to support HE projects

    • 16:00 17:30
      Covering all stage of scholarly communication with OpenAIRE services 1h 30m

      Open Science is creating a paradigm shift and is becoming the norm in research. OpenAIRE offers concrete solutions for the implementation of effective Open Science and seamless integration into the European and the global open scholarly communication ecosystem. OpenAIRE Nexus is the latest project of OpenAIRE which offers a catalogue of services in the European Open Science Cloud (EOSC) covering all stages of scholarly communication including planning, publishing, sharing, discovering, and monitoring publications, data, software, and other research products. The objective of this session is to showcase the services that support successfully embedding Open Science practices within the entirety of the research lifecycle: starting from creating a Data Management Plan with Argos to offering concrete solutions for sharing sensitive data with Amnesia, how end users can onboard data in EOSC via the OpenAIRE Provide service, and monitor the openness of the research outputs with OpenAIRE Monitor.

      The research lifecycle comprises several steps including planning, discovering and collecting research, processing data, publishing and archiving, and assessment, which are implemented and carried out using different services and tools. The services presented here are organized in different portfolios and they are an integral part of the EOSC, ensuring your compliance and use of EOSC, and are openly available to the global scholarly communication community: 

      Planning: OpenPlato, Argos 

      Discovering & Collecting: CONNECT, EXPLORE 

      Processing: Amnesia 

      Publishing: Zenodo, EpiSciences 

      Archiving: PROVIDE 

      Assessing: MONITOR, OpenAPC 

      The result of using open and sustainable services that belong to the OpenAIRE infrastructure will not only benefit researchers, but also institutions and Research Infrastructures.

      Speaker: Dr Giulia Malaguarnera (OpenAIRE)
    • 16:00 17:30
      How to evaluate and improve reproducibility? Lessons learnt and good practices with VIP and EGI 1h 30m

      Duration
      90 minutes. Constraint: the authors will only be available for giving the tutorial at the beginning of the week (Monday or Tuesday)


      Purpose
      This tutorial is aimed at all researchers interested in evaluating and improving the reproducibility of their research results. It will be based on practical examples with VIP and EGI.

      Context
      VIP is an open web portal for medical simulation and image data analysis. By effectively leveraging the computing and storage resources of the EGI federation, VIP offers its users high-level services enabling them to easily execute medical imaging applications on a large scale computing infrastructure.
      Within the ReproVIP project, we aim at evaluating and improving the reproducibility of scientific results obtained with VIP. Our final goal is to provide an integrated, end-to-end solution, allowing researchers to launch reproducible executions in a transparent manner.

      Content
      This tutorial will include a theoretical (approx. 30min) and a practical part (approx 1 hour). During the theoretical part, the audience will learn about several issues in computational reproducibility (across computing environments, libraries, versions) and how they can be handled through technical solutions.
      The practical part will focus on the tools and services provided by VIP and EGI for evaluating and improving reproducibility (e.g. Jupyter notebooks, Binder instances and the VIP API).
      - As a first step, participants will carry out reproducibility experiments on a medical imaging application. This will be done through EGI notebooks making calls to the VIP platform, where the application is available as a service.
      - The same notebooks will be used to perform analyses and visualize the variability of the application outputs between several executions.
      - Finally, these notebooks will be used to learn how any scientific work can be shared through the EGI-Replay service, by deploying an EGI-Binder instance.

      Requirements
      Participants will need to bring their own laptops. The tutorial materials will be accessible online through their web browser.

      Speakers: Gael Vila (Centre Nationale de la Recherche Scientifique (CNRS)), Mr Alexandre Cornier (Centre National de la Recherche Scientifique)
    • 16:00 17:30
      Securing Service Provider Resources with Advanced Access Management: Training

      This course is designed for those who are looking for immersive training on authorization and data protection skills essential to protect your services and resources. You will learn strategies and tools for authorization concepts, certain design patterns to solve your federated access use cases, security tokens, and how to apply EGI policy to build secure solutions using EGI Check-in. This course will include demonstrations on how to apply these features in real-world use cases. 

      Who this course is for: 

      This course is for service providers, Information Security Team members, Security Managers, Software Developers, Architects and people who are interested in the field of becoming an authorization specialist.

    • 17:30 18:30
      EGI101: Training

      Are you new to EGI? And maybe a bit confused about what EGI is, what we do, how we are structured and how we can collaborate? Then this is the session for you!
      The EGI Federation is an international e-Infrastructure set up to provide advanced computing and data analytics services for research and innovation. EGI federates compute resources, support teams and various online services from over 300 international institutes, and make those available for researchers, innovators and educators. We deliver, we enable, we innovate. This session will give a basic introduction to the EGI federation, covering all the fundamental topics that You need to know before deep-diving into the conference.
      The session will include time for Q&A as well.

      Convener: Giuseppe La Rocca (EGI.eu)
    • 18:30 19:30
      Data Spaces 101: Training

      Even for those involved, understanding "what" a data space really "is" is hard to follow. EGI has been tracking this area since the original European Strategy for Data was released in February 2020. We will share our "best" understanding of data spaces, key issues, and latest developments, hopefully in language we can all understand, to make sure our community is well informed and ready to participate!

      Convener: Mark Dietrich (EGI.eu)
    • 09:00 10:30
      From Millions to Billions: 20 years of excellence in scientific computing in Europe: Plenary

      From Millions to Billions: 20 years of excellence in scientific computing in Europe

      2023 marks the 20th anniversary of the EGI federation as European landmark infrastructure for scientific computing, an initiative conceived in the 90s that transformed forever the way infrastructures for data-intensive science are funded and operated. The EGI infrastructure of today is the largest federation of research data centres in the world, a hyperscale facility made possible by national and European investments spanning multiple regions of the world to serve more than 84,000 researchers from all scientific disciplines, and supporting a research output of more than 1,000 open access publications each year. 

      Started from DataGrid, a moonshot project funded by the European Commission and led by CERN with the participation of visionary technical architects from all over Europe, today the EGI Federation delivers cross-border access to the HTC, HPC and Cloud compute capacity of key research centres. 

      The session opens EGI 2023 and the EGI-ACE track, bringing insights into the state of play of scientific computing in Europe and internationally and its future opportunities and challenges. 

      EGI-ACE is a project that integrates capacity and data analytics tools from providers of EGI Federated Cloud and scientific communities with the European Open Science Cloud, and delivers the EOSC Compute Platform. The opening plenary and following sessions will feature how the project supported European Commission priorities and the EU Research and Innovation strategy 2020-2024.

      Welcome to EGI 2023 by Volker Guelzow, EGI Council Chair

      Welcome by Ministry of Education and Science, Marcin Michałowski, Chief Expert, Department for Innovation and Development

      Welcome from the local host PSNC by Norbert Meyer, Director of Data Processing Technologies Division at PSNC

      Scientific computing impact on open science, Tiziana Ferrari, Director, EGI Foundation

      Panel on the opportunities and challenges of scientific computing 

      Marek Margrys, CYFRONET, speaker 10’ [EGI participant, Poland]
      Alexandre Bonvin, Uni Utrecht, WeNMR PI [EGI largest user community]
      Dick M.A. Schaap, MARIS B.V. and BlueCloud {EGI-ACE user community and data provider in EOSC]
      Oxana Smirnova, Associate Professor at Lund University and NorduGrid Board Member 
      Magdalena Szwochertowska, Policy and Programme Officer @CNECT Unit C.1 Open Science and Digital Modelling at European Commission
      Marcin Michałowski, Chief Expert, Department for Innovation and Development

    • 11:00 17:30
      EGI-ACE Closing Event
      • 11:00
        Unleashing the Power of Computing in EOSC: Overcoming Challenges with Effective Solutions 1h 30m

        Nowadays science requires services that empower researchers from all disciplines to collaborate in data- and compute-intensive research across borders through free-at-the-point-of-use services. EGI is an international collaboration that federates the digital capabilities, resources and expertise of hundreds of national and international research communities in Europe and worldwide. It delivers a distributed federated infrastructure that responds to the present and future needs of data-centric scientific computing in Europe, bringing together compute facilities, data, scientific applications and software as the pillars of the EOSC Compute Platform. 

        In this session, we will present the current state of computing in EGI and EOSC and discuss its challenges and possible solutions from both user and technological perspectives. We will also talk about the EGI Federated Cloud achievements and the operational aspect of computing in EOSC while sharing the lessons learned from the EGI flagship EGI-ACE project.

        Join this session to engage with the speakers, learn from their experiences, and gain a deeper understanding of how to overcome challenges and unleash the full potential of computing within the European Open Science Cloud (EOSC). Together, we can discover effective solutions that drive scientific research and innovation to new heights.

      • 13:30
        We did it! The innovation impact of EGI-ACE 2h

        Nowadays science requires services that empower researchers from all disciplines to collaborate in data- and compute-intensive research across borders through free-at-the-point-of-use services. EGI is an international collaboration that federates the digital capabilities, resources and expertise of hundreds of national and international research communities in Europe and worldwide. It delivers a distributed federated infrastructure that responds to the present and future needs of data-centric scientific computing in Europe, bringing together compute facilities, data, scientific applications and software as the pillars of the EOSC Compute Platform. 

        The demonstration of impact is a critical aspect of scientific computing, as it allows us to evaluate the effectiveness and relevance of computational tools and methods in addressing real-research problems. To showcase the impact of scientific computing on the EGI Infrastructure and EOSC Compute platform in Europe, this conference session will feature real stories from both users and service providers of computational tools and services. These stories will illustrate how scientific computing has enabled breakthroughs in their respective domains, facilitated collaborations across borders and disciplines, and empowered researchers to address some of the most pressing challenges facing society today. Through these stories, we aim to highlight the transformative potential of scientific computing and inspire further innovation and collaboration in this field. We believe that by sharing our experiences and insights, we can foster a better understanding of the value of scientific computing and its role in shaping the future of science and technology in Europe and beyond. Users and providers will have a chance to discuss and share their perspectives in the panel.

      • 16:00
        The way forward. Computing in EOSC: evolving the role of the EGI Federation 1h 15m

        Nowadays science requires services that empower researchers from all disciplines to collaborate in data- and compute-intensive research across borders through free-at-the-point-of-use services. EGI is an international collaboration that federates the digital capabilities, resources and expertise of hundreds of national and international research communities in Europe and worldwide. It delivers a distributed federated infrastructure that responds to the present and future needs of data-centric scientific computing in Europe, bringing together compute facilities, data, scientific applications and software as the pillars of the EOSC Compute Platform. 

        As scientific computing continues to evolve and expand, it is important to consider how it can best be integrated into the European Open Science Cloud (EOSC) and what is the role of the EGI Federation and National infrastructures in this process. The session will therefore feature a panel discussion on the future of computing in EOSC and the evolving role of the EGI Federation.

        Agenda:

        Presentation (S. Andreozzi, G. Sipos): 
        Our vision for computing services in EOSC 
        Future perspectives for EGI-ACE services

        Recommendations for the stakeholders
        Panel discussion (Moderator: S. Andreozzi)

        Ignacio Blanquer, Professor at Universitat Politècnica de València (UPV) and Director at EOSC Association

        Anabela Oliveira, Head of the Information Technology Division at LNEC, Portugal

        Alexandre Bonvin, Professor at Utrecht University 

        Lukasz Dutka, Technical Director of PL-GRID, CYFRONET, Poland
        Wrap-up and final message (G. Sipos)

    • 18:00 19:00
      Opening Reception offered by EGI-ACE and ScienceMesh
    • 19:00 21:00
      Demonstrations
      • 19:20
        Demo: FedCloud client and FedCloud generic services: Dynamic DNS and Secret management 20m

        Nowadays, more and more services are dynamically deployed in Cloud environments. Usually, the services hosted on virtual machines in Cloud are accessible only via IP addresses or pre-configured hostnames given by the target Cloud providers, making it difficult to provide them with meaningful domain names. The Dynamic DNS service for EGI Infrastructure is developed for solving the problem.

        The Dynamic DNS service provides a unified, federation-wide Dynamic DNS support for VMs in EGI infrastructure. Users can register their chosen meaningful and memorable DNS hostnames in given domains (e.g. my-server.vo.fedcloud.eu) and assign to public IPs of their servers.

        By using Dynamic DNS, users can host services in EGI Cloud with their meaningful service names, can request valid server certificates in advance (critical for security) and many other advantages.

        This talk is devoted to special use cases of the Dynamic DNS service: service migration and high availability. There are many software solutions for developing high availability services but they are mostly designed for a single site or relying on load balancers. If the entire site hosting the services is down, e.g. due to power outage, software solutions like keepalived/haproxy cannot help.

        The Dynamic DNS service can be used to achieve high availability for critical services that need to operate even a whole cloud site hosting the services are down. Critical services may have backup instances deployed on other sites located on other regions to minimize the risks that all instances of the services are down at the same time. Simple scripts will check the health of instances and assign the service endpoint to a working instance via Dynamic DNS service. Implementation of such a solution via Dynamic DNS is very simple and without single point of failure. The EGI secret management service [1] is the example of the solution.

      • 19:20
        webODV - Interactive online analysis and visualization of environmental data 20m

        In the framework of the EGI-ACE project, we are deploying the webODV
        application on EGI infrastructure at
        https://webodv-egi-ace.cloud.ba.infn.it/.
        webODV is the online version of the worldwide used Ocean Data View
        Software (https://odv.awi.de). Within EGI-ACE we provide 13 ocean
        datasets (Argo, BGC-Argo, SeaDataNet) summing up to 46 GB of data, for the
        online analysis, extraction and visualization.
        webODV is easy to use and we provide video tutorials at
        https://webodv-egi-ace.cloud.ba.infn.it/webodv/videos and
        documentation at https://webodv-egi-ace.cloud.ba.infn.it/webodv/docs.
        webODV facilitates the generation of maps, surface plots, section plots,
        scatter plots, and much more for the creation of publication ready
        visualizations.
        webODV online analyses can be shared via links and are fully
        reproducible. Users can download so-called "xview" files, small XML
        files, that include all instructions for webODV to reproduce the
        analyses. These files can be shared with colleagues or attached to
        publications, supporting fully reproducible open science.

      • 19:40
        Demo: OpenRDM 20m

        Rostyslav Kuzyakiv

      • 19:40
        Demo: Structural biology in the clouds: The WeNMR portal 20m

        Structural biology aims at characterizing the structural and dynamic properties of biological macromolecules at atomic details. Gaining insight into three dimensional structures of biomolecules and their interactions is critical for understanding the vast majority of cellular processes, with direct applications in health and food sciences. Since 2010, the WeNMR project (www.wenmr.eu) has implemented numerous web-based services to facilitate the use of advanced computational tools by researchers in the field, using the high throughput computing infrastructure provided by EGI. These services have been further developed in subsequent initiatives under H2020 projects and are now operating as Thematic Services in the European Open Science Cloud (EOSC) portal (www.eosc-portal.eu).

        In this demo we will showcase the use of the WeNMR HADDOCK portal.

        Speaker: Dr Rodrigo Vargas Honorato (Utrecht University)
      • 20:00
        Demo: ENES Data Space: an EOSC-enabled and cloud-based environment for climate data analytics 20m

        The exponential increase in data volumes and complexities is causing a radical change in the scientific discovery process in several domains, including climate science. This affects the different stages of the data lifecycle, thus posing significant data management challenges in terms of data archiving, access, analysis, visualization, and sharing.

        In the context of the European Open Science Cloud (EOSC) initiative launched by the European Commission, the ENES Data Space represents a domain-specific implementation of the data space concept, a digital ecosystem supporting the climate community towards a more sustainable, effective, and FAIR use of data. The service, developed in the context of the EGI-ACE project, aims to provide an open, scalable, cloud-enabled data science environment for climate data analysis on top of the EOSC Compute Platform. It includes both ready-to-use variable-centric CMIP collections from the ESGF federated data archive and compute resources, as well as a rich ecosystem of open source modules and tools, all made available through the user-friendly Jupyter interface. The data space infrastructure has been recently enhanced with a multi-GPU node to provide accelerated computing resources, thus supporting more advanced scientific use cases, such as those based on machine learning.

        The service is accessible in the European Open Science Cloud (EOSC) through the EOSC Catalogue and Marketplace (https://marketplace.eosc-portal.eu/services/enes-data-space). It also provides a web portal (https://enesdataspace.vm.fedcloud.eu) including information, tutorials and training materials on how to get started with its main features.

        This tutorial will showcase how scientific users can benefit from the ENES data space to perform data analytics and visualization applied to climate and weather domains. More specifically, the training will cover topics from simple analytics tasks to real application examples (e.g., computation of climate indices, machine learning training and inference, etc.) using open source tools, libraries and frameworks from the Python ecosystem.

        Speaker: Fabrizio Antonio (CMCC Foundation)
      • 20:00
        Demo:Effortless service deployment on distributed cloud enviroments through the PaaS Orchestrator 20m

        The INDIGO PaaS Orchestrator, available to research communities via the EOSC Marketplace, allows users to access distributed cloud compute and storage resources in a transparent and federated manner. With the Orchestrator, users can effortlessly deploy services without the burden of locating and configuring resources.
        To further streamline the process, the Orchestrator Web Dashboard provides a collection of "pre-cooked" service templates. Once users log in, they gain access to a comprehensive service portfolio, encompassing diverse categories. Services range from instantiating virtual machines (with or without additional block storage) to automatically installing software such as Docker, Docker Compose, Elasticsearch, and Kibana, as well as deploying complex architectures like Kubernetes clusters.
        During the demo, the primary functionalities of the PaaS will be highlighted, showcasing how users can effortlessly interact with the orchestration system using both the command line interface and the user-friendly web dashboard.

        Speaker: Gioacchino Vino
      • 20:20
        Demo: Using VIP to execute medical imaging applications on EGI infrastructure 20m

        VIP (https://vip.creatis.insa-lyon.fr/) is an open web portal for medical simulation and image data analysis. By effectively leveraging the computing and storage resources of the EGI federation, VIP offers its users high-level services enabling them to easily execute medical imaging applications on a large scale computing infrastructure.
        During this demo, we will show how easy it is to use the VIP portal to launch an execution, monitor it and consult its results. We will also show how to use the large EGI infrastructure to speedup execution times in a transparent and easy way.

        Speaker: Axel Bonnet (CNRS)
    • 19:00 21:00
      Posters

      Poster Session

      • 19:40
        A digital twin engine for extreme weather events analysis on climate projections in the interTwin project 1h 5m

        Keywords: Digital Twins, Machine Learning and AI, Data management

        Climate Change has been leading to an exacerbation of Extreme Weather Events (EWEs), such as
        storms and wildfires, raising major concerns in terms of their increase of their intensity, frequency
        and duration. Detecting and predicting extreme events is challenging due to the rare occurrence of
        these events and consequently the lack of related historical data.
        Machine Learning (ML) approaches represent emerging solutions for dealing with extreme events
        analysis, providing cost-effective and fast-computing methods that can complement or replace
        traditional methodologies. Such solutions require huge amounts of heterogeneous data for properly
        training and running the models, which in turn pose big challenges in terms of data management,
        computing/memory resource requirements, workflow orchestration and software infrastructure
        needs. A Digital Twin for EWEs integrating data and models could provide scientists and policy
        makers with a system for conducting prompt analysis and evaluating what-if scenarios.
        In the context of the EU-funded interTwin project, a Digital Twin for the analysis of extreme events,
        targeting tropical cyclones and wildfires, on future climate projections following a data-driven
        approach is being developed. The interTwin project aims at defining a Digital Twin Engine for
        supporting scientific Digital Twins applications from different fields. This contribution will present
        the initial work behind the design of this machine learning-powered Digital Twin for extreme events
        studies as well as some preliminary results.

      • 19:40
        AI-based Policy Making Process 1h 5m

        Public policy development is becoming more and more challenging for public administrations and the recent events (first of all the Covid-19 Outbreak) demonstrated how an evidence-based approach is crucial for managing critical situation with short responses, fast adaptation and citizens’ support.

        The ultimate vision of data-driven policy making entails the use of Artificial Intelligence (AI) as a means of increasing the efficiency of the policy development and management (i.e. going beyond development to adaptation and optimization) process and boosting a more responsive, adaptive, intelligent and citizen-centric governance.

        Cross-industry standard process for data mining, known as CRISP-DM, is the most widely-used open standard process model that describes common approaches used by AI experts.
        AI4PublicPolicy leverages this model to promote an AI-based Policy Making process addressing the objectives below:

        • Provide a data-driven, AI-based and evidence-based approach to
          policymaking
        • Promote and facilitate the collaboration between Policy Makers and AI
          Experts
        • Involve citizens and other stakeholders in the Policy evaluation and
          optimization
        • Boost the acceptance of the policies presenting and explaining the
          Policy development outcomes
        • Reuse and share Policy development models and datasets

        The main steps of the AI4PublicPolicy AI-based Policy Making Process are as follows:

        • Phase 1 - Policy Definition: In this step the Policy Maker starts creating an Analytical Policy Model describing the Policy problem(s)
          and associating and describing relevant Datasets for the Policy.
        • Phase 2 - Policy Extraction: In this step an AI Expert, or the Policy Maker him/herself with AutoML support, creates one or more AI Workflows to analyse the datasets and prepare the data to train and test AI Models based on different AI Algorithms, in order to provide responses (insight, recommendations) to the Policy problem(s). The Policy Maker executes the AI Models on new data, analyses the responses and validates the AI Models.
        • Phase 3 - Policy Evaluation: In this step the Policy Maker involves the relevant stakeholders (citizens, business and other local actors). in the Policy Evaluation creating one or more surveys on the Policy problems, AI model responses and Policy alternatives. The Policy Maker then evaluates stakeholders’ feedback, presented with a statistical or sentiment scoring, and decides to complete the Policy with actionable outcomes or optimizing it taking into account stakeholder’s feedback.
        • Phase 4 - Policy Presentation and Sharing: In this step the Policy Maker uses eXplainable AI (XAI) techniques provided by the platform to better understand and explain the rationale behind the AI Models responses. Once completed this step the Policy Maker could present the final Analytical Policy Model, which represent the result of the AI-based Policy Making Process, to the relevant stakeholders and publish it in a shared catalogue.
        Speaker: Alessandro Amicone (GFT Italy)
      • 19:40
        Blue-Cloud 2023 1h 5m
        Speaker: Dick Schaap (Mariene Informatie Service MARIS BV)
      • 19:40
        Building OpenStack Cloud with Kubernetes as Undercloud 1h 5m

        OpenStack is an open-source cloud computing platform that provides a comprehensive set of services for building and managing private, public, and hybrid clouds. However, deploying and maintaining an OpenStack cloud can be a complex and time-consuming process. In recent years, Kubernetes has emerged as a popular container orchestration platform, making it an attractive option for managing the underlying infrastructure of OpenStack.

        At CESNET we have decided to build a next-generation of OpenStack cloud utilizing Kubernetes. In this poster, we will present our new architecture for building the CESNET OpenStack cloud site using Kubernetes as the undercloud with OpenStack Helm infra based distribution. We will discuss benefits of this approach which enables simplified deployment and management of OpenStack services by leveraging the automation and scalability of Kubernetes.

        Speaker: Mr Adrian Rosinec (CESNET)
      • 19:40
        C-SCALE Workflow Solutions for Earth system monitoring, modelling, and forecasting 1h 5m

        C-SCALE Workflow Solutions presents a suite of open-source, reusable components designed to facilitate the development and deployment of Earth system monitoring, modelling, and forecasting applications on FedEarthData. These components are derived from various C-SCALE Case Studies, with solutions ranging from hydrodynamic and water quality modelling to real-time satellite-derived surface water area estimates. Users can leverage these pre-built templates to customize their applications for specific applications and geographic areas of interest.

        Key features of the C-SCALE Workflow Solutions include data downloading and preparation, model execution, post-processing, and visualization, including analysing data in an interactive Jupyter Notebook environment. By utilizing these readily available components, users can efficiently create tailored workflows for their specific needs, capitalizing on the extensive resources and infrastructure provided by C-SCALE's Cloud and HPC platforms.

        The C-SCALE community repositories host these open-source solutions, allowing for seamless deployment on FedEarthData, which offers uniform access to a federation of computing and data providers for Copernicus and Earth Observation workloads. The diverse range of solutions currently available, as well as those under development, demonstrates the potential for further expansion and customization to support Earth system monitoring, modelling, and forecasting applications across various domains.

      • 19:40
        Cloudification of scientific experimental data using Onedata 1h 5m

        Scientific research produces increasingly large amounts of experimental data that can be challenging to manage, store, and share. To address these challenges, we developed software that automates acquiring, management, and sharing of data produced by specialized devices. Our software is built on top of Onedata –⁠ a data management solution used by the established EGI DataHub service, providing a reliable and scalable platform for managing large datasets and seamless integration with other processing and data management systems.

        One of the key features is its ability to automatically make experimental data available to the scientific community in a predefined and accessible way, particularly useful for on-the-fly processing in local or distant data centers, real-time analysis, or archiving to permanent storage according to defined quality of service (e.g., data distribution).

        Speaker: Tomáš Svoboda (Masaryk University)
      • 19:40
        Computational services in support of Coastal Digital Twins 1h 5m

        Digital Twins integrate continuously, in an interactive, two-way data connection, the physical and the virtual world. They provide a virtual representation of a physical asset enabled through data and models and can be used for multiple applications such as real-time forecast, system optimization, monitoring and controlling, and support enhanced decision making. The development of Digital Twins, targeting user creation of knowledge and products, along with the development of global initiatives for virtual representation of the Earth (DestinEarth) and oceans in particular (Digital Twins of the Ocean - EDITO model lab and Iliad projects) have set the stage for a user-centered digital vision for support of all coastal and ocean interventions and knowledge creation.
        Actions within Digital Twins are materialized through computational services, devoted to address specific concerns, such as operational prediction of water dynamics. The concept of web-based coastal computational services, available through user-friendly interfaces, have promoted their usefulness for coastal knowledge creation and coastal management.
        Herein, we showcase three computational services for coastal dynamics developed at LNEC and LIP: OPENCoastS (https://opencoasts.ncg.ingrid.pt/), for on-demand forecasting of coastal dynamics, WORSICA (https://worsica.incd.pt/), for Sentinel2 image processing for inland and coastal water bodies delimitation, and jUMP (http://jump-app.lnec.pt/index/), for simulation of underwater noise propagation. Three applications of these services will presented aiming at illustrating their importance for creating reliable and accurate services within Coastal Digital Twins.

      • 19:40
        Easy Deployment of Dask and Jupyter Notebooks on Managed Kubernetes? 1h 5m

        Data analytics platforms have become increasingly important in modern research,
        enabling scientists to analyze large datasets and extract valuable insights.
        Reproducible and open science is a crucial aspect of nowadays scientific research, ensuring
        that results are transparent, verifiable, and can be reproduced by other
        researchers. Workflow tools, Virtual Research Environments (VREs), and
        computational notebooks are among the many tools that may help with this goal.
        However, deploying these tools can be complex, time-consuming, and also not
        successful in the end.

        We provide a managed Kubernetes platform that can be used for reliable and
        easy deployment. In this poster, we present several use cases including
        Jupyter notebooks, Binder deployment with Kaniko builder, Dask
        compute clusters, and S3 backend storage as success story of services
        built on the managed Kubernetes platform.

        Speaker: Mr Adrian Rosinec (CESNET)
      • 19:40
        ELKH Cloud: milestones towards EOSC and ESFRI 1h 5m

        The federated science cloud of the Eötvös Loránd Research Network, ELKH Cloud is one of the award-winner research infrastructures in Hungary. Members of the scientific community are not only using but are also developing and operating the cloud services: the Institute for Computer Science and Control (SZTAKI) and the Wigner Research Centre for Physics provide the computing and data services to more than 250 research projects since 2016 (the inauguration with the support from the Hungarian Academy of Sciences).
        Based on positive feedback received in recent years, as well as growing demand for artificial intelligence applications, the cloud capacity was significantly expanded by 2022 with support from ELKH. As a result, 7344 vCPU, 72 GPUs, 35 TB RAM, 338 TB SSD storage, 1.25 PB HDD storage and 100 Gbps network capacities have become available to the users.
        Research often requires complex, large-scale platforms based on the coordinated operation of multiple components. The ELKH Cloud therefore provides customisable, reliable and scalable reference architecture templates, among others with the help of cloud orchestration methods.
        In addition to operating systems and basic IaaS level cloud functionalities, the most popular artificial intelligence research and data ingestion frameworks are also available at PaaS level, and recently a quantum computing related reference architecture has been released. The enhanced ELKH Cloud provides a competitive research infrastructure that also welcomes projects initiated by universities and national laboratories.
        ELKH aims to make the enhanced ELKH Cloud an integral part of the European Open Science Cloud (in the EGI-ACE project) and the SLICES ESFRI initiatives (in the SLICES-SC and SLICES-PP projects).

      • 19:40
        Empowering European researchers with federated Earth Observation data analytics: An overview of the C-SCALE services 1h 5m

        The H2020 Copernicus-eoSC Analytics Engine (C-SCALE) project provides European researchers with a federated Earth Observation (EO) data analytics platform by leveraging the best-of-breed tools and services of pan-European e-infrastructures. C-SCALE delivers a seamless user experience by abstracting away the complexity of resource provisioning and orchestration. The project offers three main services: the Federated Earth System Simulation and Data Processing Platform (FedEarthData), the Earth Observation Metadata Query Service (EO-MQS), and the openEO Platform service. FedEarthData provides a distributed infrastructure of data and compute providers to support the execution of EO workflows at scale, while the EO-MQS makes Copernicus data distributed across partners discoverable and searchable. Finally, the openEO platform simplifies processing and data management by offering intuitive programming libraries alongside a large EO data repository. This poster provides an overview of the C-SCALE services and their benefits to European service providers and researchers.

      • 19:40
        EUreka3D - European Union’s REKonstructed content in 3D 1h 5m

        EUreka3D is a project funded by the Digital Europe Programme of the European Union, to support the digital transformation of the cultural heritage sector, by offering capacity building and training, and new services, to Cultural Heritage Institutions facing the challenge of advancing in the digitization effort, especially in 3D digitization, access, storage, sharing.

        The project will offer a capacity building and knowledge programme, next to services and resources developed in a piloting action based on smart technical infrastructures and tools, also registered on the European Open Science Cloud (EOSC). This virtual space will allow cultural heritage institutions to use storage and computing resources to manage their 3D assets, which will give smaller institutions an affordable and simpler way to digitise, model, manage and share 3D records.

        Additionally, the project will aggregate and publish a variety of high-quality 2D and 3D digital records in Europeana.eu website as well as offer clear examples for the use and reuse of the cultural content in neighbouring areas such as tourism and education.

        Finally, it will support institutions by developing training material for 3D digitisation, data processing and publication on Europeana.eu as well as organising training activities to support cultural heritage institutions with the use of the virtual space.

        Website: https://eureka3d.eu

      • 19:40
        FitSM in the age of digital transformation 1h 5m

        FitSM is a lightweight standard for IT service management. It brings order and traceability with simple, practical support and provides a common conceptual and process model setting out realistic requirements.

        With FitSM Training you will learn the fundamentals of IT service management and how to implement FitSM in your organisation through a combination of lessons and examples. FitSM professional training is certified by internationally recognized authorities.

        While FitSM is a TRL9, it is proven in the operational environment.
        With FitSM training you can:
        - Increase your expertise in managing IT services
        - Raise your professional profile with a recognised certification

        Analysing FitSM use cases can be a powerful resource for helping businesses understand their opportunities and how to use FitSM in addressing business needs related to IT service management, while a use case is a detailed description of how a user interacts with a system or product to achieve a specific goal.

      • 19:40
        Green Computing progress and improvements within EGI-ACE - Poster 1h 5m

        Task T7.2 of EGI-ACE was formulated with the key intention of making EGI-ACE greener. In order to achieve this the task’s working group focused upon:

        • Discovering existing good practice already being undertaken by the
          project partners
        • Knowledge dissemination amongst partners(particularly in relation to
          the existing good practice)
        • Measuring key metrics to illustrate progress relating to Green
          computing over the lifetime the project

        Activities undertaken with these aims in mind were:

        • Survey
        • Taskforce
        • Signposting
        • Webinars

        The working group produced a report “D7.4 - Green Computing progress and improvements within EGI-ACE” which covered the activities, results of the activities and a number of key observations, conclusions and recommendations:

        • There are pockets of good practice amongst the partners.
        • Overall engagement amongst partners is low
        • A multi-faceted approach, combining reward and compulsion alongside
          enabling activities/policies is recommended to combat this. Broadly speaking this could be expressed as:
          • Build upon the current knowledge sharing activities
          • Facilitate the adoption of green initiatives and practices
          • Incentivise the adoption of green initiatives and practices
          • De-incentivise non green activities and non adoption of green initiatives and practices

        This poster highlights some of the observations, conclusions and recommendations along with the methods used to formulate them.

      • 19:40
        How to adopt FAIR and Open Science principles in EOSC: supporting the research lifecycle management within Earth Science communities, with the RELIANCE project services 1h 5m

        The RELIANCE Project (Research Lifecycle Management for Earth Science Communities and Copernicus users in EOSC) aims to realize the vision of FAIR research in EOSC and enable the adoption of Open Science practices in EOSC by implementing an holistic research management approach based on three key and complementary technologies: i) Research objects (RO) as the overarching mechanism to manage scientific research activities, which relies upon ROHub platform as the reference service; ii) data cubes as the mechanisms enabling an efficient and scalable Earth Observation data access and discovery, which relies upon the Advanced geospatial Data Management (ADAM) platform as the reference service; iii) text mining and semantic enrichment services allowing to extract machine-readable metadata from RO resources, enabling researchers to discover scientific information at scale and to structure their own research, and which rely on the AI-based platform COGITO as base system. As part of the integration in EOSC, RELIANCE services leverage and integrate with some of the EOSC core-cross cutting and added value services, playing a complementary role to what is already available and bridging between various EOSC services.
        RELIANCE validated the project services through multidisciplinary and thematic real life use cases led by three different Earth Science communities: Geohazards, Sea Monitoring and Climate Change communities. In our presentation, we will showcase different types of scenarios for the three Earth Science communities represented in Reliance to highlight how the scientists in our respective disciplines fostered their work towards Open Science.

      • 19:40
        iMagine 1h 5m
        Speaker: Ilaria Fava (EGI.eu)
      • 19:40
        OpenCitations: an open science infrastructure on its way towards sustainability 1h 5m

        OpenCitations is a community-based open science infrastructure organization with the mission of harvesting and openly publishing accurate and comprehensive metadata describing the world's academic publications and the scholarly citations that link them, under open licenses at zero cost and without restriction for third-party analysis and re-use. This means that OpenCitations’ activities generate no income to sustain its operations. Currently, OpenCitations operates on a mixed funding model that includes voluntary effort from its Directors, in-kind support from the University of Bologna, financial contributions from supporting organizations including donors and members that are used to support staff and computational infrastructure, and time-limited grant income from public and private funding agencies including the EU that is used solely for time-limited research and development projects. However, despite the valuable support from the community and the successful involvement in the SCOSS second funding cycle, OpenCitations is still seeking a ‘business model’ which could help move it from being a 'sustainable infrastructure' (in POSI terms) to being a financially sustained infrastructure. To strengthen its weak points, OpenCitations has recently launched two internal Working Groups dedicated to the themes of “Governance Evolution” and “Community Building”, which in OpenCitations are deeply connected. Indeed, OpenCitations’ stakeholder community is integrated into OpenCitations’ existence both in a financial way, by providing economic support via membership and donations, and strategically, by guiding activities in the OpenCitations Council (made up of OpenCitations members) and the elected International Advisory Board. Moreover, OpenCitations never stops to build connections with other Open Science actors and initiatives, and to look for external advice as an active member of the “SCOSS family” group and the “POSI adopters” group. OpenCitations activities are community-driven and community-dedicated. For this reason, OpenCitations’ way towards sustainability needs collective guidance and effort with the mission to make OpenCitations a solid and enduring infrastructure upon which the global scholarly community itself can rely for open bibliographic and citation metadata for many years to come.

        Speaker: Chiara Di Giambattista (University of Bologna - OpenCitations)
      • 19:40
        The IBiSCo Project 1h 5m

        The IBiSCo Project (Infrastructure for BIg data and Scientific COmputing) is an initiative founded by the Italian Government within the National Operational Plan (PON) 2014/2020 to create a large e-infrastructure in South Southern Italy, contributing to the development of the "Pillar 2: Infrastructure" of the "Important Project of Common Interest on High Performance Computing and Big Data enabled Applications" (IPCEI-HPC-BDA).
        IBiSCo involves some of the main Italian research institutes: INFN, CNR, INAF, INGV and the University of Napoli Federico II and University of Bari Aldo Moro. The project made possible to acquire a large amount of resources distributed over the sites of Bari, Catania, Napoli and Frascati, interconnected over thenetwork of Research and Education network.
        The resulting infrastructure is empowered with a geographically federated HTC, HPC and Cloud resources and it is ready to be expanded in the context of PNRR, representing a pillar of the new Italian initiatives supporting e-Science and Innovation.
        In this poster we will present the status of the e-infrastructure implementation, the achievements of the project after more than 3 year of activity (both in terms of active services and user communities supported) and the plans to advance at national and international level.

        Speaker: Silvio Pardi (INFN)
      • 19:40
        Towards Coastal Digital Twins for bathing waters protection: predicting water quality with OPENCoastS+ 1h 5m

        Digital Twins can support decision-making in bathing waters management by providing virtual representations of physical assets that can be used for real-time forecast, system optimization and monitoring and controlling. Herein, the OPENCoastS+ service is demonstrated in the coastal region of Albufeira (Portugal), as a stepping stone to the implementation of a Digital Twin for bathing waters protection.
        Stormwater discharges from the city of Albufeira and its upstream peri-urban area can contaminate the adjacent bathing waters. Real-time tools are thus required to predict the quality of bathing waters and support the assessment of the need to prohibit beach water usage.
        OPENCoastS+ (https://opencoasts.ncg.ingrid.pt) is a service that builds on-demand hydrodynamic and water quality forecast systems for user-selected areas and maintains them running operationally. The service generates forecasts of water levels, 2D or 3D velocities, temperature, salinity, waves and water quality variables, including Feacal Indicator Bacteria - FIB (Escherichia coli and enterococcus). The relevant physical, chemical and biological processes are simulated using the modeling suite SCHISM and several forcing options are available for the ocean, river and surface boundaries. The service is provided through the European Open Science Cloud (EOSC) computational resources.
        The OPENCoastS+ service was implemented in the coastal area of Albufeira. All the relevant forcings (ocean, streams, urban drainage and atmosphere) are included in the model. The service provides daily forecasts of 3D circulation (water levels, coupled wave-currents, salinity, and temperature), FIB concentrations and associated indicators for the classification of the bathing waters quality.

        Speaker: Marta Rodrigues (LNEC - Laboratório Nacional de Engenharia Civil)
      • 19:40
        Towards digital twin oriented Virtual Research Environment 1h 5m

        LTER-LIFE is a large-scale research infrastructure in the making; it aims to provide a state-of-the-art e-infrastructure to study and predict how changes in climate and other human-induced pressures affect ecosystems and biodiversity. One of the grand challenges in ecology is to understand and predict how ecosystems are impacted by changes in environmental conditions, and external pressures. Predicting how complex ecosystems will behave under different scenarios requires combining empirical and observational data. To truly predict how ecosystems will respond to current and future global change, we need to increase the availability of existing environmental long-term datasets, integrate disparate types of ecological data, and create a user-friendly and secure cloud-based digital modelling and simulation platform that can be used to link data to models and scenarios. This can be done by creating Digital Twins of entire ecosystems. A Digital Twin is “a digital replica of a living or non-living physical entity”. Building Digital Twins of ecosystems has only recently become possible as Big Data, artificial intelligence (AI) applications and analytics, advanced computing infrastructure, and the FAIR principles have been developed and made available for ecology, ecosystem restoration, and biodiversity science. This approach brings together data scientists, informaticians, and ecologists in research communities around ecosystems. Detailed information on the LTER-LIFE project can be found at https://lter-life.nl/.

        The Virtual Laboratories provide VRE components, data, models, and other Big Data tools (e.g., AI) and allow scientific users to discover, access, and integrate research assets for specific scientific purposes and manage the deployment, execution, and provenance of in silico experiments (analysis and simulations) on local or remote platforms. This includes the assets: 1) findable, accessible, interoperable, and reusable (i.e. FAIR) datasets of long-term studies of plants and animal populations as well as of environmental data, 2) FAIR models such as (a) process-based models that formalise ecological knowledge on abiotic and biotic relationships, or (b) data-driven models that explore relationships between data using classical statistics, or more contemporary artificial intelligent tools. These FAIR models can be discovered, accessed, and reused in new scientific workflows. 3) Rules of interactions between assets (i.e., data and models), including solutions for scaling and enhancing interoperability of datasets, which are collected with different methodology. And 4) Tools for scenario studies to explore the effects of in situ local management strategies on ecosystem functioning. These assets can be coupled in various ways (which is enabled by the LTER-LIFE platform) depending on the objectives of the specific research. Together, they will ultimately compose a Digital Twin of a specific ecosystem, which can then be used for scientific research or scenario studies to address societal issues.

        With LTER-LIFE, we will offer a single point of entry for scientists studying entire ecosystems or components.
        The LTER-LIFE portal gives users tailored access to assets and facilities. LTER-LIFE develops: 1) Data bases and process models that are FAIR, by applying (meta)data standards, controlled vocabularies, ontologies, and persistent identifiers. 2) Workflows that allow coupling of these data and models at the desired temporal and spatial scale, and that provide Big Data methodology to analyse these data (AI and machine learning approaches such as deep learning and active learning which enable new theory discovery and novel insights from high-dimensional sparse data spaces) as well as tools for uncertainty analysis, scenario studies and forecasting. 3) Virtual Labs that build upon infrastructure services and components of a VRE, allowing interdisciplinary communities of scientists to collaborate via a digital platform (including Big Data storage, cloud-based modelling, and dynamic simulations). And 4) Basic and advanced training in the use of the infrastructure.

        LTER-LIFE will be an open-source infrastructure and users can either use the developed assets within their own local environment or run it in the cloud, facilitated by LTER-LIFE through our partner SURF. The development will focus on well-selected scientific use cases from the Veluwe and Wadden Sea ecosystems. The development of LTER-LIFE will be incremental and follow Agile practices. Setting up the LSRI LTER-LIFE as an open-source infrastructure aligns well with the current FAIR and Open Science developments. New data or models can be added as assets, which can then be used by other users. Workflows and data as used in projects will receive a digital identifier and will be stored by LTER-LIFE’s partners DANS-KNAW and SURF.

    • 09:00 10:30
      The computing and data continuum in Europe: state of the art and the user community perspective: Plenary

      The amount of data gathered, shared and processed in frontier research is set to increase steeply in the coming decade, leading to unprecedented data processing, simulation and analysis needs. Research communities are preparing to launch new instruments that require data and compute infrastructures several orders of magnitude larger than what is currently available and entering in the Exascale era. To meet these requirements, new data-intensive architectures, heterogeneous resource federation models, and IT frameworks will be needed, including large-scale compute and storage capacity to be procured and made accessible at the pan-European level. Additionally, the emergence of high-end Exascale HPC and Quantum computing systems provides new opportunities for accelerating discoveries and complementing the capabilities of existing research HTC and Cloud facilities. Addressing key questions around scalability, performance, energy efficiency, portability, interoperability and cybersecurity is crucial to ensuring the successful integration of these heterogeneous systems. 

      Objectives

      Understand the needs of data-intensive scientific collaborations to have access to a European Exabyte-scale research data federation and compute continuum.
      Learn about existing initiatives, the current state of the art and open challenges
      Understand how research infrastructures and e-Infrastructures can jointly address common research and innovation needs

      Convenor: Tiziana Ferrari, EGI Foundation

      10‘ Welcome to session by the convenor - T. Ferrari, EGI Foundation

      30’ The Destination Earth initiative, Thomas Geenen, ECMWF

      40’ Panel: Requirements and state of the art for data, HTC, HPC and Cloud federation in Europe

      Thomas Geenen, Technology Partnership Lead at ECMWF
      Maria Girone, CERN openlab Head and CERN-IT EC projects Lead
      Luciano Gaido, Director of Technology at INFN
      Andrej Filipcic, IIS
      Andrea Manzi, Data Solutions Manager at EGI Foundation
      5’ Concluding remarks

    • 10:30 11:00
      Demonstrations
      • 10:30
        EGI & C-SCALE: Notebooks for Earth Observation 20m

        The C-SCALE project has been federating compute and data resource providers around centralized EGI services, aiming at providing users with seamless access to processing capacities as well as source data for their analyses. Alongside the traditional IaaS and PaaS services, Jupyter Notebooks have been identified as an environment suitable not only for interactive analysis within C-SCALE, but also for documenting the different steps one needs to take in discovering and accessing geospatial data across Europe. The demonstration of C-SCALE's example notebooks and procedures will focus on those essential features: simple steps to get started using the federated resources for interactive resources of Earth observation data.

        Speaker: Petr Pospisil
      • 10:30
        EGI Workload Manager Service: activities in the EGI-ACE project 20m

        The EGI Workload Manager Service (WMS) is based on the DIRAC interware, offering a complete software solution for communities of users that need to exploit distributed, heterogeneous computing and storage resources for big data processing. This presentation will give an overview of the service and then focus on the WMS functionalities supporting the current trends and technology innovations in the framework of the EGI-ACE project. A dedicated training session is foreseen during the conference for the interested users.

        Speaker: Dr Gino Marchetti (CNRS / CC-IN2P3)
    • 11:00 13:00
      Data analytics platforms, tools and VREs for EOSC 2h

      Over the last years the vast amounts of data created by fundamental research domains force us to consider new solutions for capturing, managing and processing of this data. Extracting knowledge or useful insights from this data deluge can be used for smart decision-making in various scientific domains. In the area of data science, a data analytics platform is an ecosystem of services that allow scientific researchers and Virtual Research Environments to perform big data analytics. A data analytics platform is an ecosystem of services that allow scientific researchers and Virtual Research Environments (VREs) to perform big data analytics. In this track we present how Data Analytics platforms, tools and VREs are helping scientific communities to daily work of modern scientific communities, Research Infrastructures, and the central role that EGI services is playing in this regard.

    • 11:00 12:30
      National infrastructures for Open Science in EGI and beyond 1h 30m

      EGI is an aggregator of national and domain specific capabilities, with its governance based on national e-infrastructures (NGIs) and international scientific communities. The EGI federation brings together compute & storage resources, scientific datasets and various digital platforms and data analysis and simulation environments from over 30 countries, 300 institutes to make these available for Open Science worldwide. 

      This session will zoom into the NGI, the national pillars of EGI, and into similar national providers of the global compute ecosystem. The session will include a series of talks that (1) highlight new services and plans from EGI national stakeholders, and (2) demonstrate alignment pathways for national EGI nodes and the ‘European Open Science Cloud’.

      Agenda:

      Robert Lovas (SZTAKI, Hungary): ELKH Cloud - milestones towards EGI, EOSC and ESFRI
      Giacinto Donvito (INFN, Italy): INFN-Cloud: a national distributed and federated cloud infrastructure supporting scientific communities
      Jerome Pansanel (CNRS): EGI and Open Science activities and plans in France
      Matthias Schramm (TU Wien, Austria): Emerging solutions for national and international EGI and EOSC users in Austria
      Jerry Horgan (Walton Institute, Ireland): Experiences and plans for compute delivery for Irish and EGI communities
      Eric Yen (Academia Sinica, Taiwan): Asian collaboration on disaster mitigation based on deeper understanding approach
      Sa-kwang Song (KISTI, South Korea): Korea Research Data Commons - Focusing on Research Data and Software

    • 11:00 12:30
      OpenStack: Experience, challenges , solutions to OpenStack operations in the EGI Federation 1h 30m

      EGI Cloud federates more than 25 OpenStack cloud providers into a multi-cloud IaaS with Single Sign-On, application and data sharing and common operations for supporting the computing needs of research communities.Each of these providers operates its own OpenStack deployment with complete freedom on the internal architecture and uses its preferred tools for configuration management and monitoring.

      This session aims to share the challenges and solutions for operating highly-available, large-scale OpenStack deployments from current EGI Cloud providers and commercial providers.

      Subsessions

      Integrating EGI Check-in with OpenStack alongside other Open ID Connect providers.
      11:05 AM - 11:15 AM
      Speaker: Tom Clark

      Beskar Cloud: OpenStack deployment on top of Kubernetes
      11:20 AM - 11:35 AM
      Speaker: Adrián Rošinec

      CloudFerro. OpenStack and Identity Federation
      11:35 AM - 11:45 AM
      Speaker: Paweł Turkowski

      Location: Room: Barcelona

      Speakers: Enol Fernandez (EGI.eu), Gianni Dalla Torre (EGI.eu), Tom Clark (EODC GmbH), Mr Adrian Rosinec (CESNET), Paweł Turkowski (CloudFerro)
    • 11:00 12:30
      Unlocking the Potential of Copernicus Data: Explore the C-SCALE Service Offer for Advanced Earth Observation Analytics
    • 11:00 12:30
      Vision, business model and sustainability of the major European Data initiatives 1h 30m

      The session  aims to bring together leading experts from these initiatives to discuss the critical aspects of governing and sustaining data spaces in the digital age. The event will focus on sharing experiences and best practices for effective governance of data spaces and ensuring their long-term sustainability.

      Key topics that will be covered during the event include data privacy, data sovereignty, interoperability, and business models for data exchange. 

      Participants will have the opportunity to engage in insightful discussions with experts from different organizations and learn about the latest trends, challenges, and opportunities in the field. Overall, this session promises to be a valuable platform for sharing knowledge and promoting collaboration among stakeholders involved in data space governance and sustainability.

      Part 1: EOSC+ AI on Demand + Data Spaces: Vision, business models and sustainability of the major EU initiatives 

      EOSC, Ignacio Blanquer , EOSC Association, 10 min
      Data Spaces, Daniel Alonso, DSSC,  10 min
      AIoD, Barry O Sullivan UCC, AI4Europe Coordinator, 10 min (remotely)
      Presentation of an EDIC in Agricultural DS - Sebastien Picardat, 5min

      Panel

      Moderated talk,  Marta Gutierrez  30 min

      Speakers in the stage: Ignacio Blanquer, Daniel Alonso, Gabriel Gonzalez

      Round table:

      EDIC, the new buzzword. Is an EDIC an appropriate structure for these initiatives? Other experiences /  recommendations in the creation of legal entities ( ERICs, AISBLs, Non for Profit…) 
      What information do you need to collect to sharpen your understanding of the customer segment (including their willingness to pay)?
      Overlapping or complementarity? Room for collaboration and user navigation between the 3 initiatives
      How can EGI help? Role of EGI as e-infrastructure and service provider in supporting these initiatives
      What obstacles would the customer segments face in working with these initiatives?

    • 12:30 13:30
      Dataspaces and AI/ML services in the agricultural domain 1h

      In this session we will present several ongoing initiatives that are related with creation of the Agriculture dataspaces. We will present as well the selected AI/ML focused projects that will contribute to the Dataspaces

      Introduction to the dataspaces in Agriculture -  AgriDataSpace project (Raul Palma - PSNC)
      Building the Agriculture data Space in Poland - eDWIN and Datamite project (Maciej Zacharczuk - WODR)
      Building the AI Testing Experimental Facilities in agriculture and contribution to agriculture dataspaces - AgriFoodTEF project (Łukasz Łowiński - PIT)
      Agriculture AI based use cases in the ICOS and AI4EOSC projects(Marcin Plociennik - PSNC)
      Platform for yield and cost prediction of agricultural production - PRAGMATIC (Tomasz Wojciechowski - PULS)

    • 13:00 14:00
      Demonstrations
      • 13:00
        DigiTwin 3D – a data-driven 3D interactive view on wind farms in the North Sea 20m

        Effective use of data driven intelligence for complex planning and decision making requires innovations that enable stakeholders to better understand the type of information that the data systems provide. In many cases stakeholders have limited expertise on a specific technical subject, but still need to understand and interpret the data driven intelligence to be able to act on limitations, consequences and alternatives.
        Three-dimensional data visualization in a virtual or augmented web environment can be such an innovation that helps to interpret data by emerging the user in a virtual world where the data is visualized realistically. An example of such an environment is a planned offshore wind farm, where the user is able to move and look around freely to examine energy yields, effects on fisheries, shipping industries and ecology.
        At MARIS we have developed a demo digital twin of a wind farm in the North Sea with the goal to determine to what extend 3D visualization can provide insight to data and assist marine spatial planning in the North Sea. The demo allows for the user to freely move around in the Prinses Amalia wind farm that is made to scale. The wind farm contains multiple assets that the user can investigate for more information and includes real time information of the water and air temperature and wind speed and direction.

        Speaker: Tjerk Krijger (MARIS)
      • 13:00
        pDNSSOC, deploying SOC capabilities in under 1 minute 20m

        In the Research & Education (R&E) sector, “defending as a global community” is a crucial strategy. This involves having the ability to produce and share relevant threat intelligence, and the capability to leverage it accross the entire sector.

        Large campuses and research labs typically have dedicated security teams and adequate skills. But even there, deploying and operating a Security Operations Center (SOC) is a hard problem to solve.

        In addition, a significant part of the R&E sector has a relatively low maturity level on threat intel, and no or limited security staff and expertise.

        pDNSSOC provides a turn-key solution to both detect and respond to security incidents.

        Speaker: Pau Cutrina Vilalta
      • 13:20
        Recommender Metrics Framework (RMF): Measuring the success of a Recommender System 20m

        The Recommender Metrics Framework (RMF) is an independent "metrics framework as a service" that supports the evaluation and adaptation of recommendation mechanisms. The European Open Science Cloud (EOSC) uses a modern recommender System (RS) in the EOSC Marketplace for suggesting various resources. Using RMF to measure the success of the EOSC RS is crucial to get valuable insights into many aspects that affect the user experience. The use of additional diagnostic metrics and visualizations offers deeper and sometimes surprising insights into a model's performance. The evaluation is quantitatively performed by processing information such as resources, user actions, ratings, and recommendations to measure the impact of the AI-enhanced services and user satisfaction as well as to incorporate this feedback and improve the services provided, via a user-friendly Application Programming Interface (API) and a User Interface (UI) dashboard. The framework supports real-time ingestion of data, multiple resource types and recommendation systems as sources, while it consists of 3 components. The Preprocessor component is responsible for: data retrieval through a connector module that claims and transforms data from various sources, leveraging item-associated knowledge, dissociated and dummy data removal, relation tags dispatch to information that marks various associations in the data, i.e. registered or anonymous-related users and services, statistics' information provision. Moreover, this component is able to run in batch mode for gathering data from the various sources offline, and in stream mode where agents gather the various data from the respective sources in real-time. The RSmetrics component is responsible for processing the collected data, computing the designated evaluation metrics, and producing the necessary information in a homogenized manner. The third component is a RESTful API along with a rich UI dashboard presenting reports as a web service and visualizing metrics and statistics. The current version of the implementation features simple statistics and complex metrics. Particularly, it delivers the metrics: Catalogue and User Coverage, Diversity based on the Gini Index and the Shannon Entropy, Novelty, Accuracy, and a list of Key Performance Indicators (KPI)s indicating measurable values that demonstrate how effectively key business objectives are achieved. Concerning the latter case, the KPIs metrics are: Click-Through Rate, Hit Rate, Top-5 ordered and recommended Services, Top-5 categories and scientific domains for both orders and recommendations. The software not only delivers the results through the data REST API and the UI dashboard, but it additionally exposes graphic visualizations of statistics such as User Actions and Recommended Items per Day and Month. The RS evaluation framework is constantly expanding with new features, metrics, and utilities, to lead to more robust, data-adaptable, and good-quality RS designs. In this demo we are going to present the way the different components work, the integration with other EOSC Components, a typical process flow, and an overview of the statistics, the metrics, the KPIs, and the graphs produced after the computations.

        Speaker: Mr Nikolaos Triantafyllis (National Infrastructures for Research and Technology (GRNET))
      • 13:20
        Step-Up Authentication in the Perun Identity and Access Management System 20m

        Multi-factor authentication (MFA) is becoming a standard feature of Authorization and Authentication Infrastructures (AAIs). Alongside the other security measures, it allows accessing services and datasets securely.

        Not only the access to datasets and services needs to follow the security rules. An Identity and Access Management (IAM) system as a place where the access to resources is managed has to implement the same practices. Otherwise, the attacker could exploit the weaker part of the access chain in the identity and access management process instead of overcoming highly secured datasets and services access.

        To address the security risks, we have focused on implementing the MFA into the Perun IAM system. The challenge was to create a well-adapted solution that balances security, usability, and accessibility. It led us to implement a step-up authentication process triggered only when users want to manage access to sensitive resources inside the system.

        This demonstration will go through the challenges we encountered during the design and implementation and showcase the final solution in the Perun IAM system.

    • 14:00 16:00
      Advanced Cloud Services for Research 2h

      This session showcases several initiatives that deliver and expand cloud services for research communities: new providers, tools for improved access to federated providers, access to new computing paradigms and platforms for reproducibility and large-scale analytics.

    • 14:00 15:40
      Emerging Architectural Directions for Federated Digital Infrastructures 1h 40m

      Recognizing ongoing work on innovative architectures for federated digital infrastructures, this session will present key architectural directions in GAIA-X, Data Spaces Business Alliance, W3C, CERN  and EOSC. 

      Objectives 

      To find potential areas of alignment or common architectural components, what is successful, what does not work.

      To find gaps or challenges that still need to be addressed by architectures to solve complex challenges supporting big data for AI/ML, complex research environments, IoT etc… 

      Will standardisation help in the road to convergence ?  

      Do particular deployment models favour federations in any particular way  ? Blockchain, centralised, decentralised, wallets, pods etc … 

      Session talks: presentations of major architectural families (60')

      Technical Convergence efforts of the Data Spaces Business Alliance 12’ Daniel Alonso

      GAIA-X architectural evolution 12’ Pierre Gronlier (remote)

      Foundational Elements for Federated Digital Infrastructure from the W3C 12’, Rigo Wenning (remote)

      Sharing applications and data between institutions with Open Cloud Mesh  Hugo Gonzalez Labrador 12’

      To EOSC and Beyond: Future Directions of the European Open Science Cloud, Diego Scardaci 12’ 

      Discussion & Panellist: moderated by Mark Dietrich (30')

      Daniel Alonso BDVA
      Rigo Wenning,  W3C
      Pierre Gronlier,  GAIA-X
      Hugo Gonzalez Labrador  & Giuseppe Lo Presti CERN
      Diego Scardaci EGI

    • 14:00 15:00
      Green Computing 1h

      This session will host presentations and discussions about recent activities and achievements on lowering the environmental impact of computing and analysis of Big Data within EGI Federation. It will include findings of the EGI-ACE Green Computing Task Force - good practice and knowledge dissemination undertaken by project partners, also will showcase the solutions taken to reduce the environmental impact of a Cloud site part of EGI Federated Cloud.

      Last but not the least, based on results and achieved experience, a potential roadmap to go further ‘green’ will be recommended to be implemented across data centres within the Federation, including how data centres can be influenced to support the adoption of ‘Green IT’ across their estate.

    • 14:00 17:00
      Shaping services for the EOSC: the experience of EOSC Future and the infraEOSC 07 projects 3h

      Over the past two and half years, EOSC Future and five other H2020 projects, C-SCALE, DICE, EGI-ACE, OpenAIRE Nexus, and Reliance, have been working and collaborating to enrich EOSC with new services and resources from different national, regional and institutional infrastructures across Europe. As the projects are approaching their final mile, it is time to gather to reflect on the enhanced EOSC-Exchange capabilities and their benefits to scientific communities. This session is divided into two parts which are open to all but have slightly different angles. First, we will hear presentations on the services developed in these projects, which might be especially interesting for other current or potential EOSC service providers. The latter part of the session focuses on real-life use cases of these services. This panel discussion might be especially interesting for anyone interested in using EOSC services in the future.

    • 14:00 18:00
      Trust, Security and Identity: 4h

      Solutions and use cases for federated trust and identity management aiming at trusted environments to exchange and analyse sensitive data

    • 14:00 18:30
      WLCG Community Meeting (Grid Deployment Board - GDB) 4h 30m

      The Grid Deployment Board (GDB) is part of the Worldwide LHC Computing Grid Collaboration (WLCG) and meets every month. Normally those meetings are held at CERN, exceptionally in other places.This session will host the June GDB meeting.

      The topics under discussion concerns operational aspects affecting the WLCG community and the EGI Infrastructure.

      The agenda can be consulted in: https://indico.cern.ch/event/1225113/

      The Zoom connection details are shown in the indico page (if logged in)

      14:00 → 14:10: Introduction (J. Flix)

      14:10 → 14:45: Tier-1 deployment at NCBJ (H. Giemza)

      14:45 → 15:10: UMD update plans (J. Pina)

      15:10 → 15:30: Status report on sites migrating away from DPM (A. Paolini)

      15:30 → 16:00: EOS Workshop summary (O. Keeble)

      16:00 → 16:30: Coffee break

      16:30 → 17:00: Using EGI Check-in tokens (V. Ardizzone)

      17:00 → 17:30: Status of the new Helpdesk (P. Weber)

      17:30 → 18:00: ARC update and future outlook (M. Pedersen)

      18:00 → 18:30: Cloud Services for Synchronisation and Sharing (CS3) 2023 summary (O. Keeble)

    • 15:30 16:00
      Demonstrations
      • 15:30
        Infrastructure Manager (IM): TOSCA deployment and orchestration in the Cloud Continuum 20m

        The Infrastructure Manager (IM) is a TOSCA-based orchestrator that has been developed throughout the last decade to support the deployment of complex application architectures across a myriad of IaaS Cloud back-ends, including the most popular public Cloud Platforms (Amazon Web Services, Microsoft Azure, Google Cloud Platform) as well as European public providers such as Open Telekom Cloud, Cloud & Heat and Orange Cloud, European federated platforms such as EGI Cloud Compute and on-premises platforms like OpenStack or OpenNebula. Furthermore, the IM has been running in production for more than 5 years in EGI Cloud Compute, where it is the preferred solution to deploy complex infrastructures in the cloud.

        This contribution demonstrates a TOSCA extension, to define FaaS-based applications to be deployed using open-source serverless platforms. Also, the IM has been extended to support application deployment described using these new TOSCA templates. For this, we have relied on the serverless capabilities of the two existing open-source developments. On the one hand, SCAR allows the deployment of Docker-based applications on AWS Lambda and AWS Batch. On the other hand, OSCAR can run event-driven Docker-based applications for data processing purposes on Kubernetes clusters, including low-powered ARM-based devices such as Raspberry PIs clusters, thus approaching the edge of the cloud continuum scenario.

        Speaker: Miguel Caballer (Universitat Politècnica de Valencia)
      • 15:30
        VISA at DESY - scientific analysis tools from your browser 20m

        The recently finished EU projects PaNOSC and ExPaNDS produced a scientific analysis portal called VISA (Virtual Infrastructure for Scientific Analysis) that is now being used by many Photon and Neutron facilities in Europe. Originally developed by the Institut Laue-Langevin, VISA has been adopted by many project partners and is currently being developed further as a community project in order to ensure a sustainable subsequent use in the future.
        The portal is envisioned to facilitate authorized user access to experiment data directly from a web browser. For this purpose, the portal manages virtual machines on which experiment data is mounted, which can either originate from a recent experiment or from a FAIR data repository. Furthermore, the portal administrators at the respective institue hosting a VISA instance can provide a variety of analysis tools that fit the scientists' use cases and preferences.
        During the talk, the portal and its functionalities will be presented as well as the current deployment strategy at DESY.

        Speaker: Dr Tim Wetzel
    • 15:40 16:20
      Artificial Intelligence and Machine Learning: Lightning Talks 40m

      This lighting talk session will provide a glimpse into the innovative ways Artificial Intelligent and Machine Learning technologies, from cloud to edge, are being used to address real-world challenges in different fields, from smart cities to biology.

    • 16:30 18:30
      Artificial Intelligence and Machine Learning in distributed/federated environments 2h

      Jointly organised by AI4EOSC, AI4Europe and iMagine

      This session will feature several projects working on Artificial Intelligent and Machine Learning and contributing to the European Open Science Cloud (EOSC) and AI-on-Demand (AIoD) platform. It will begin with an overview of these projects, providing insights into their architecture, services, and applications. Then, in the second part of the session, the focus will shift towards interoperability and integration among the projects, with an open discussion around technical and architectural issues, identifying common interest areas and sharing experiences, to foster a collaborative environment aimed at advancing the development and implementation of the AI platforms and services.

      In the workshop part of this session we will cover the major initiatives around the development and implementation of artificial intelligence and machine learning services in the European Open Science Cloud. We will discuss technical and architectural issues, identifying common interest areas and sharing experiences. Relevant use cases will also be invited, in order to showcase their experiences and expectations.

    • 17:00 18:00
      Towards a sustainable funding for EOSC 1h

      This session will briefly present the results of the consultation survey of the EOSC Association Financial Sustainability Task Force which collected the EOSC community’s views on proposals for the financial sustainability of EOSC. Key members of the EGI community will discuss the results’ relevance to and alignment with business models for EGI.

      The session will discuss topics including the EOSC Core, EOSC Exchange and the federation of data in EOSC.

      Participants will have the opportunity to engage in discussion about EGI's business models and their alignment with potential EOSC business models, consequences for EGI of choices made for EOSC business models, and consequences for EOSC of EGI's business models.

      The survey was developed to ask the EOSC community about their views on the proposals for the financial sustainability of the main components of EOSC - EOSC Core, EOSC Exchange and Data Federation in EOSC - described in the Financial Sustainability Task Force’s interim progress report Towards Sustainable Funding Models for the European Open Science Cloud, published in November 2022.

    • 19:00 23:00
      Conference Dinner
    • 09:00 10:30
      Making Space for Data

      Data is – and has been – fundamental to research.  The Research Data Alliance (RDA) was formed in 2013 (10 years ago!) and now has almost 13,000 members from around the world.  The seminal “FAIR Guiding Principles for scientific data management and stewardship” paper was published in 2016 and galvanised the open data movement in science.

    • 10:30 11:00
      Demonstrations
      • 10:30
        AI Models as a Service with OSCAR: Integration with dCache and EGI Notebooks 20m

        OSCAR is an open-source platform built on Kubernetes for serverless event-driven data-processing applications. To trigger the execution of these applications, OSCAR receives events from object-storage providers such as MinIO.

        An OSCAR cluster can be easily deployed on the EGI Federated Cloud, as well as many other public and on-premises Clouds, through the Infrastructure Manager (IM), both integrated in the EOSC Portal.

        Our past contribution demonstrated the capabilities of OSCAR, through the execution of different use cases, to create event-driven data-processing workflows along the computing continuum, processing part of the workload on the Edge and delegating the compute-intensive processing on a public cloud such as EGI Federated Cloud and Amazon Web Services (AWS).

        In this contribution, we showcase, on the one hand, the integration of OSCAR with the distributed system for scientific data storage dCache, using it as an event source to trigger the execution of services. Using Apache Nifi to manage the event ingestion workflow between dCache and OSCAR, we have created a use case involving AI/ML models from the Deep Open Catalog for scalable asynchronous inference.

        Moreover, we implemented a new Python API to trigger OSCAR services from Jupyter Notebooks, such as EGI Notebooks. We prepared another demo from EGI Notebooks to showcase interactive synchronous inference of AI/ML models.

        Project PDC2021-120844-I00 funded by MCIN/AEI/10.13039/501100011033 funded by the European Union NextGenerationEU/PRTR. Also, grant PID2020-113126RB-I00 funded by MCIN/AEI/10.13039/501100011033. This work was supported by the project “interTwin’’ which has received funding from the European Union’s Horizon Europe Programme under Grant 101058386. Also, by the project AI4EOSC ‘‘Artificial Intelligence for the European Open Science Cloud’’ which has received funding from the European Union’s Horizon Europe Research and Innovation Programme under Grant 101058593

        Speakers: Caterina Alarcon Marin (Universitat Politècnica de València), German Molto (Universitat Politècnica de València)
      • 10:30
        Demo: openEO Platform - federated computational infrastructure for Earth Observation Data 20m

        openEO Platform is an operational, federated service including backend infrastructure at EODC, VITO and Sinergise which enables users to access and process large amounts of Earth Observation data. This service advances the knowledge about planet earth and provides users with a range of advantages including the ability to process data fast and share analysis with other users.

        During this demonstration session, we will showcase how to use the platform via Python Jupyter Notebooks and a graphical user interface. Attendees will gain insight on how to sign up, sign in, submit first small jobs and gain an introduction to larger scale EO processing.

    • 11:00 12:00
      EGI Digital Innovation Hub - Delivering value to the private sector 1h

      EGI DIH is the Digital Innovation Hub created by EGI in 2022 to stimulate the collaboration of EGI with the private sector, formalising years of experiences supporting companies in innovation activities.The EGI DIH offers companies the possibility to join its community via different levels of engagement and where valuable services at technical, networking, training and funding level will be facilitated as part of the value proposition of the EGI DIH. 

      The session will present the EGI DIH to the audience, with the participation of one of the success stories where we will learn the experience of the Finish company Binare IO working with the EGI DIH. Later Aruba, an active collaborator of EGI, will participate as an invited speaker presenting its commercial cloud service to the community.

    • 11:00 13:00
      Public Authorities: EOSC Support, Engagement and Exploitation of Services 2h
    • 11:00 18:00
      Science Mesh Community meeting

      The CS3MESH4EOSC final event, entitled "Science Mesh - Unlocking Open Science and Collaborative Research Landscape" will showcase how the Science Mesh is contributing to an easier and more robust open science across Europe, thanks to novel approaches for data sharing and synchronisation.

      Detailed agenda here: https://cs3mesh4eosc.eu/news-events/events/cs3mesh4eosc-final-event-science-mesh-unlocking-open-science-and-collaborative
      The Science Mesh is a federation of distributed data storage and sharing services, known as Enterprise File Sync and Share (EFSS) services, where cloud platforms from different providers are interconnected. Science Mesh users can easily share and manage data through the familiar interfaces of their institutional services and collaborate with colleagues from other organisations. The Science Mesh services are not specific to any particular research discipline, since it addresses the needs of a wide array of users interested in scientific collaboration. It can be used either by user communities within existing Research infrastructures such as European Research Infrastructure Consortia (ERICs), National Research and Education Network (NRENs) and other official clusters, but also by individual researchers and smaller research groups. All this without compromising data privacy and security, with the advantage of making data within the Science Mesh FAIR.

      The first half of the event will count with live demonstrations, where each data service from the Science Mesh will be presented from a user-perspective point of view. Event attendees will get practical information on how they can join the Science Mesh as a researcher, a software developer or a service provider. The event will also bring together representatives of Science Mesh use cases, who will explain how Science Mesh is making a difference in their lives, thanks to easier data sharing and synchronisation. A panel discussion with representatives of different sectors, from research to industry and education, will discuss the most urgent trends & priorities for cross-border science collaboration between different sciences.

      The second half of the event will be focused on the technical novelties within the Science technical foundation. A series of demonstrations will be presented, followed by a panel discussion, with representatives of CS3MESH4EOSC members that are part of the EOSC Task Forces, on how the Science Mesh contributes to EOSC’s success and the overall EOSC Strategic Research and Innovation Agenda (SRIA).

      The event will close with an overview of future Science Mesh developments, as well as how the tool will become self-sustained after the project's conclusion.

    • 11:00 12:00
      Trusted Environments for the Processing of Sensitive Data 1h

      Trusted Research Environments (TREs) enable researchers to have access to data and resources in a safe way. These environments contain various requirements even if there are also a number of similarities because of the nature of the sensitive data.

      Main motivation within this session is to explore different solutions and architectures within service providers. Additionally, the session focuses on EGI services and also discusses possibilities to create predefined cross border TRE service within the EGI community for project purposes. 

      Program:

      Introduction
      A Secure Data Infrastructure that Supports Data Visiting; Martin Weise, Andreas Rauber (TU Wien)
      Lightning Talks on technical solutions and architectures of TREs
      Secure Processing Environments and Secure Data Archives in the Nordics and DICE-EOSC, Abdulrahman Azab (University of Oslo)
      In-situ processing of sensitive medical imaging data - The CHAIMELEON repository; Lozano Lloret Pau, Andy S. Alic, Sergio López Huguet, José Damián Segrelles Quilis, Ignacio Blanquer
      LETHE architecture solutions and lessons to learn; Ville Tenhunen
      Wrap-up

    • 13:00 14:00
      Demonstrations
      • 13:00
        A Reference Architecture for Data Analytics for Autonomous Systems 20m

        In this demonstration we would like to present the improvements to our data analytics platform for autonomous systems [5]. It is a scalable, cloud-agnostic and fault-tolerant data analytics platform we built using open-source components composed as reusable building blocks [3]. It is used as a reference architecture in different research projects such as DIGITbrain [1] and within the National Laboratory for Autonomous Systems [2] (abbreviated as ARNL in Hungarian) and within the TKP2021-NVA-01 project in Hungary focusing on hydrogen-powered, cooperative and autonomous remote sensing devices and connected data processing framework research. The main use cases of our reference architecture include the areas of smart/autonomous production systems (collaborative robotic assembly) and autonomous vehicles (mobile robots with smart vehicle control). The improved capabilities of the platform include among others: (a) support for native Robot Operating System (ROS) based, Internet-of-Things (IoT) based and generic time-series (Apache Kafka) based data collection; (b) a native cloud-agnostic software container-based architecture; and (c) support for open science research data via uploading and sharing datasets via ELKH ARP [4].

        Furthermore, we would like to demonstrate the capabilities of our reference architecture via a use case from ARNL: hardware-in-loop smart vehicle control [6]. This use case provides a hierarchical control for automated vehicles with which their safe and efficient motion in roundabouts can be guaranteed. The control hierarchy contains vehicle level and cloud level. The goal of the cloud level is to achieve enhanced control performances using the high computation capacity of the cloud. Thus, reinforcement learning on the cloud level for achieving maximum speed of the vehicles is implemented. Moreover, on the vehicle level the safety requirement, i.e., collision avoidance, is guaranteed. The advantage of the solution is that safe performance specifications even at the degradation of the communication in the network can be guaranteed. Significant novel content of this work is the implementation of the method using indoor test vehicle environment with cloud connection. [5]

        Our demonstration will focus on (i) the custom use case and platform components, (ii) the data collection from the indoor test vehicle environment with smart vehicle control and (iii) the visualization capabilities of the platform at the different stages of the data collection and processing.

        This work received funding from the European Union within the framework of the National Laboratory for Autonomous Systems (RRF-2.3.1-21-2022-00002). Also this work was funded by European Union’s Horizon 2020 project titled ”Digital twins bringing agility and innovation to manufacturing SMEs, by empowering a network of DIHs with an integrated digital platform that enables Manufacturing as a Service (MaaS)” (DIGITbrain) under grant agreement no. 952071. Project no. TKP2021-NVA-01 has been implemented with the support provided by the Ministry of Innovation and Technology of Hungary from the National Research, Development and Innovation Fund, financed under the TKP2021-NVA funding scheme.

        References
        [1] DIGITbrain H2020 project (2020). https://digitbrain.eu/.
        [2] National Laboratory for Autonomous Systems (NLAS/ ARNL). https://autonom.nemzetilabor.hu .
        [3] A. Cs. Marosi et al., "Toward Reference Architectures: A Cloud-Agnostic Data Analytics Platform Empowering Autonomous Systems," in IEEE Access, vol. 10, pp. 60658-60673, 2022, doi: 10.1109/ACCESS.2022.3180365.
        [4] ELKH ARP - CONCORDA (Concentrated Cooperation on Research Data). https://science-data.hu .
        [5] A. Cs. Marosi et al., “Towards Reference Architectures: A Cloud-agnostic Data Analytics Platform Empowering Autonomous Systems” EGI Conference 2022 Demonstration. https://indico.egi.eu/event/5882/contributions/16700/
        [6] Balázs Németh and Péter Gáspár. The design of performance guaranteed autonomous vehicle control for optimal motion in unsignalized intersections. Applied Sciences, 11(8), 2021.

        Speaker: Mr Márk Benjamin Emődi (SZTAKI)
      • 13:00
        Improving UX of the Account Linking Process 20m

        The number of digital identities possessed by a single user significantly increased over the past years. The problem is that users want to access services regardless of the identity they use. Account linking is a process of joining users' identities that comes forward as a solution to the problem.

        However, even today, the whole process is often handled either manually by an administrator or using a user-driven approach that is typically too complicated for users to navigate through smoothly. Users need a straightforward interface that will guide them step-by-step through the whole process. This is especially important when you need to support less technically skilled users, as we discovered during onboarding research communities to Life Science AAI.

        Fortunately, using the OIDC protocol, we can easily integrate the account linking directly to existing web applications. This design leads to a secure self-service solution with the desired improvement of the overall user experience (UX).

        This demonstration introduces challenges in the account linking process addressed in the Life Science AAI. Moreover, it describes how they can be addressed using the OIDC protocol and demonstrates the integration with other applications. Most of the experience and the solution itself are generic and can be used even outside the Life Science domain.

      • 13:35
        Demo: Onedata - workflow processing, distributed datasets and archives 20m

        During the demo, we will showcase the latest features of Onedata, introduced in version 21.02.1.

        1. Multi-cloud data processing using automation workflows and archive preservation.
        2. Managing the overall lifecycle of distributed datasets, from preparation, through annotation and dissemination, to archiving for long-term preservation.

        The demo will be performed on the Onedata services at EGI DataHub, with the intention of easy reproducibility by EGI users. Onedata version 21.02.1 is being gradually introduced to DataHub since May.


        Onedata [1] is a high-performance data management system with a distributed, global infrastructure that enables users to access storage resources worldwide. It supports various use cases ranging from personal data management to data-intensive scientific computations. Onedata has a fully distributed architecture that facilitates the creation of a hybrid-cloud infrastructure with private and commercial cloud resources. Users can collaborate, share, and publish data, as well as perform high-performance computations on distributed data using POSIX-compliant data access applications.

        The Onedata ecosystem comprises several services, including Onezone, which is the authorisation and distributed metadata management component that provides access to the system. Oneprovider delivers actual data to the users and exposes storage systems to Onedata, while Oneclient enables transparent POSIX-compatible data access on user nodes. Oneprovider instances can be deployed as single nodes or HPC clusters, on top of high-performance parallel storage solutions that can serve petabytes of data with GB/s throughput.

        The latest Onedata release version, 21.02.1, introduces the integration of a powerful workflow execution engine, which is powered by OpenFaas [2]. This integration enables the creation of complex data processing pipelines that can leverage the transparent access to distributed data provided by Onedata. The workflow functionality is especially useful for creating a comprehensive, OAIS compliant data archiving and preservation system, which covers all archival requirements, including ingestion, validation, curation, storage, and publication. The workflow function library includes ready-to-use functionalities, which are implemented as Docker images and cover typical archiving actions such as metadata extraction, format conversion, and checksum validation. Additionally, new custom functions can be easily added and shared among user groups. The solution underwent thorough testing on auto-scalable Kubernetes clusters, ensuring its reliability and scalability. Since Onedata is a distributed solution, deploying multiple instances of Oneprovider and OpenFaas on separate clouds provides a transparent data-plane layer for multi-cloud workflow processing.

        Moreover, version 21.02.1 introduces several new features and improvements that enhance Onedata's capabilities in managing distributed datasets throughout their lifecycle. The software allows users to establish a hierarchical structure of datasets, control multi-site replication and distribution using Quality-of-Service rules, and keep track of the dataset size statistics over time. In addition, it also supports the annotation of datasets with metadata, which is crucial for organizing and searching for specific data. The platform also includes robust protection mechanisms that prevent data and metadata modification, ensuring the integrity of the dataset in its final stage of preparation. Another key feature of Onedata is its ability to archive datasets for long-term preservation, enabling organizations to retain critical data for future use. This is especially useful in fields such as scientific research, where datasets are often used for extended periods or cited in academic papers. Finally, Onedata supports data-sharing mechanisms aligned with the idea of Open Data, such as the OAI-PMH protocol and the newly introduced Space Marketplace. These features enable users to easily share their datasets with others, either openly or through controlled access.

        Currently, Onedata is used in European EGI-ACE [3], PRACE-6IP [4], and FINDR [5] project, where it provides a data transparency layer for managing large, distributed datasets on dynamic hybrid cloud containerised environments.

        Acknowledgments: This work was supported in part by 2018-2020’s research funds in the scope of the co-financed international projects framework (project no. 5145/H2020/2020/2).

        [1] Onedata project website. https://onedata.org.
        [2] OpenFaaS - Serverless Functions Made Simple. https://www.openfaas.com/.
        [3] EGI-ACE: Advanced Computing for EOSC. https://www.egi.eu/projects/egi-ace/.
        [4] Partnership for Advanced Computing in Europe - Sixth Implementation Phase. http://www.prace-ri.eu.
        [5] FINDR: Fast and Intuitive Data Retrieval for Earth Observation.

        Speaker: Lukasz Dutka (CYFRONET)
      • 13:35
        HEP Benchmark Suite: the centralized future of WLCG benchmarking 20m

        When dealing with benchmarking, result collection and sharing is as essential as the production of the data itself. A central source of information allows for data comparison, visualization, and analysis, which the community both contributes to and profits from.

        While this is the case in other fields, in the High Energy Physics (HEP) benchmarking community both script and result sharing required human interaction in the past, which greatly hindered data collection and retention.

        Thus, in the context of the HEP Benchmark Project, a tool has been devised to unify the benchmark execution process as well as the format of the reported results: the HEP Benchmark Suite. The Suite is complemented by an OpenSearch instance, allowing for additional functionalities such as long-term storage, visualization, and monitoring.

        This state-of-the-art solution aims at solving this challenge present in the HEP community by becoming its benchmarking cornerstone for all HEP sites around the globe. Its architectural design will be described in this contribution, together with the experience acquired while gathering thousands of measurements worldwide across the WLCG.

    • 13:00 18:00
      interTwin Tech F2F
    • 14:00 15:00
      Plenary

      As we draw the EGI Conference 2023 to a close, join us for a forward-looking keynote address on the future of computing. 

      Sagar Dolas, Program Manager at SURF Innovation Labs, will enlight us on the future of computing & its impact on grand challenges in society. The keynote will be followed by an interview to dive into how the main trends relates to the service strategy of the EGI Federation.

      The last part of the session will be devoted to final remarks, ackowledgments and to announce winners of conference competitions.

      Don't miss this opportunity to look towards the future of computing while also reflecting on the collective efforts that made the EGI Conference 2023 an unforgettable experience.

    • 15:00 16:30
      EOSC Platform - Highlighting EGI contribution to EOSC Future 1h 30m
    • 15:00 16:30
      EuroScienceGateway: an open infrastructure for data-driven research 1h 30m

      The EuroScienceGateway (ESG) project, in collaboration with the European Open Science Cloud (EOSC) and Galaxy, aims to provide a user-friendly scientific workflow and data integration platform. This initiative is focused on creating an open infrastructure for data-driven research that is robust, scalable, and integrated, allowing for federated computational infrastructures and democratization of research data analysis. The ESG project involves 17 European partners who are working towards integrating metadata and knowledge from EOSC services like Zenodo and Workflowhub.eu, to enhance the FAIRification of data analysis.
      In our session, we will provide an introduction to Galaxy, a platform that is increasingly being used by scientific communities. Specifically, we will take a closer look at usegalaxy.eu, the largest European instance, and discuss how it is administered and continuously developed. Our discussion will cover recent data developments such as object stores and distributed computing, which are relevant to a variety of research communities. We will explore principles and solutions related to open data initiatives like iRODS or Rucio and highlight challenges faced by new communities such as material sciences and astrophysics that are exploring Galaxy as part of ESG. Additionally, we will focus on the development of Pulsar (our distributed compute network), a core component of ESG, and our efforts to mature it to TLR-9, enabling the efficient utilization of resources across Europe.

    • 15:15 16:45
      Data Spaces: governance consultation 1h 30m

      Data Spaces, as called by the European Strategy for Data, present a multi-stakeholder governance challenge.

      The GREAT project is defining the governance framework underlying the Green Deal Data Space.
      The defition process involves co-design with the community, validation and endorsement to ensure a successful implementation.
      This session will present the Governance framework and gather feedback and suggestions for improvements from the community in an interactive way.

      An example of some of the complexities surrounding the governance of data spaces includes data monetisation. The Datamite project will present an overview of the challenges and solutions ahead in this area.

    • 15:15 16:45
      Federation management solutions - Latest achievements and technical roadmaps 1h 30m

      This session will present the status and technical roadmaps of the ‘Federation management solutions’ that the EGI federation relies on to integrate distributed resources and services into a Pan-European and Global infrastructure. 

      The session will include short talks that present the recently added new functionalities of these solutions, and their future development plans for the coming years. 

      The portfolio of EGI ‘Federation management solutions’ include: 

      Accounting Repository and Accounting portal to collect HTC, Cloud, and Storage records and display them with different types of aggregation

      A Configuration Database (GOCDB) storing the EGI Infrastructure topology information

      A Messaging Service powered by ARGO Messaging Service (AMS), a real-time messaging service that allows users to send and receive messages between independent applications.

      A Monitoring Service, based on ARGO, to check the status of the EGI Infrastructure and create performance reports

      An Helpdesk Service based on GGUS to record incidents and service requests and to perform support activity to users and providers

      An Operations Portal for supporting the operations and coordination of the EGI Infrastructure.

      A Software Provisioning Infrastructure providing quality verified middleware to the EGI federated providers.

      A Security Coordination service managing and carrying on activities concerning EGI Infrastructure policies, Software Vulnerabilities, and security incidents handling.

    • 16:30 17:45
      Polish Data repositories for EOSC 1h 15m

      Presents use cases of European and national data repositories which
      use the research infrastructure of National Data Storage (NDS) in Poland, among others biology, digital humanities, astrophysics data.
      NDS provides a platform for open science repositories supporting the EOSC approach in Europe. 

      The session will include 4 presentations of application use cases from
      different science disciplines and presentation of the data services provided via NDS platform.

    • 09:00 10:30
      Bringing your AI models to EOSC 1h 30m

      Tutorials:
      Introduction to the AI4EOSC platform (ca.10’)
      Build AI application /service based on existing AI modules using the AI4EOSC platform (ca.25’)
      Develop a new AI application / service with the AI4EOSC platform (ca.25’)
      Please, follow Preparation Steps BEFORE the session!
      Deploy your AI-based service for inference using the AI4EOSC platform (ca.25’)
      Speakers:
      Ignacio Heredia (CSIC), Khadijeh Alibabaei (KIT), Amanda Calatrava (UPV) (maybe: Valentin Kozlov (KIT))

      Description:
      The AI4EOSC delivers an enhanced set of advanced services for the development and serving of Artificial Intelligence (AI), Machine Learning (ML) and Deep Learning (DL) models and applications in the European Open Science Cloud (EOSC). The project builds its technology based on the results of the DEEP-Hybrid-DataCloud project and other previous ICT projects, where the partners have developed new innovative services to support machine learning over the cloud. In AI4EOSC, we evolve those services to better support the new challenges ahead when building distributed AI applications and improve scalability and performance for large-scale AI applications..In the hands-on workshop, the AI practitioners and all interested people will learn how the AI4EOSC/DEEP AI platform supports typical scenarios for AI-based applications and services: re-use of existing AI modules on the marketplace, AI model development from scratch and training, and model serving.

      Acknowledgment: AI4EOSC receives funding from the European Union’s Horizon Europe 2022 research and innovation programme under agreement 101058593

      Relevant links:
      AI4EOSC project: https://ai4eosc.eu
      AI4EOSC/DEEP platform documentation: https://docs.deep-hybrid-datacloud.eu
      OSCAR framework for FaaS: https://docs.oscar.grycap.net/
      Jupyter notebook repository (for the "Develop AI application" part): https://git.scc.kit.edu/m-team/ai/ai4eosc-egi23-tutorial

    • 09:00 10:30
      ENES Data Space: an EOSC-enabled and cloud-based environment for climate data analytics 1h 30m

      The exponential increase in data volumes and complexities is causing a radical change in the scientific discovery process in several domains, including climate science. This affects the different stages of the data lifecycle, thus posing significant data management challenges in terms of data archiving, access, analysis, visualization, and sharing.

      In the context of the European Open Science Cloud (EOSC) initiative launched by the European Commission, the ENES Data Space represents a domain-specific implementation of the data space concept, a digital ecosystem supporting the climate community towards a more sustainable, effective, and FAIR use of data. The service, developed in the context of the EGI-ACE project, aims to provide an open, scalable, cloud-enabled data science environment for climate data analysis on top of the EOSC Compute Platform. It includes both ready-to-use variable-centric CMIP collections from the ESGF federated data archive and compute resources, as well as a rich ecosystem of open source modules and tools, all made available through the user-friendly Jupyter interface. The data space infrastructure has been recently enhanced with a multi-GPU node to provide accelerated computing resources, thus supporting more advanced scientific use cases, such as those based on machine learning.

      The service is accessible in the European Open Science Cloud (EOSC) through the EOSC Catalogue and Marketplace (https://marketplace.eosc-portal.eu/services/enes-data-space). It also provides a web portal (https://enesdataspace.vm.fedcloud.eu) including information, tutorials and training materials on how to get started with its main features.

      This tutorial will showcase how scientific users can benefit from the ENES data space to perform data analytics and visualization applied to climate and weather domains. More specifically, the training will cover topics from simple analytics tasks to real application examples (e.g., computation of climate indices, machine learning training and inference, etc.) using open source tools, libraries and frameworks from the Python ecosystem.

      Speaker: Mr Fabrizio Antonio (Centro Euro-Mediterraneo sui Cambiamenti Climatici (CMCC))
    • 09:00 10:30
      Get experience with computing on Kubernetes 1h 30m

      Containers are becoming more widespread in the IT industry, they offer the advantage of having the entire software and runtime environment packaged together in a single package. The deployment of the containerized application might be tricky, it requires to have a container orchestrator installed - therefore the user has to usually manage the entire virtual machine with all required system services just to run the container.
      In recent years, it has been shown how the adoption of the popular container orchestrator - Kubernetes has made the work of developers and system administrators more efficient. The developers or users of computing resources do not have to debug the operating system, secure it and take care of individual system services just to run, for example, a web application or a database.

      Join our workshop and gain hands-on experience with computing on Kubernetes, and how it might help you with your research activities. Together, we will discuss the benefits and limitations of managed Kubernetes service. Then we will get our hands dirty going through the deploying of the secure application and exposure to the Internet.

    • 09:00 10:30
      Onedata hands-on workshop: Workflow Processing, Dataset Management and Archiving 1h 30m

      Onedata [1] is a high-performance data management system with a distributed, global infrastructure that enables users to access storage resources worldwide. It supports various use cases ranging from personal data management to data-intensive scientific computations. Onedata has a fully distributed architecture that facilitates the creation of a hybrid-cloud infrastructure with private and commercial cloud resources. Users can collaborate, share, and publish data, as well as perform high-performance computations on distributed data using POSIX-compliant data access applications. The latest Onedata release introduces the integration of a powerful workflow execution engine, which is powered by OpenFaas [2]. This integration enables the creation of complex data processing pipelines that can leverage transparent access to organizationally distributed data. In addition, the new software version offers several new features and improvements that enhance its capabilities in managing distributed datasets throughout their lifecycle.

      This hands-on workshop will focus on the latest Onedata release version, 21.02.1. Participants will explore its features through interactive exercises, with a special focus on data processing using automation workflows, distributed dataset management, and archive preservation. Other covered topics will include directory size statistics and the Space Marketplace. The training materials will correspond to the scenarios from the Onedata demonstration, presented during another session. The workshop will be conducted on the Onedata services at EGI DataHub, with the intention of easy reproducibility by EGI users.

      Keywords: workflows, distributed dataset, data lifecycle, data access, distributed systems, file sharing.

      Acknowledgments: This work was supported in part by 2018-2020’s research funds in the scope of the co-financed international projects framework (project no. 5145/H2020/2020/2).

      [1] Onedata project website. https://onedata.org.
      [2] OpenFaaS - Serverless Functions Made Simple. https://www.openfaas.com/.

      Speaker: Lukasz Opiola (CYFRONET)
    • 09:00 16:00
      Security Training:

      Threat intelligence and Security Operations Centres
      In the current research and education environment, the threat from cybersecurity attacks is acute, having grown in recent years. We must collaborate as a community to defend and protect ourselves. Efficient collaboration and response require both the use of detailed, timely and accurate threat intelligence alongside fine-grained networking monitoring. In this session, we explore aspects both of sharing appropriate intelligence and the conceptual design of a security operations centre, including recent work towards a SOC environment appropriate for cloud infrastructures.

      Security of OIDC deployments
      We can see services, users, and infrastructures migrating to OIDC or already using the technology. We will welcome site and services administrators, VO managers and users, developers, and security experts. We want to trigger discussions about the security implications of the new technology, the level of readiness, changes in the habits of users, and the overall impact on security operations and incident response.
      SSC Forensics Walkthrough
      We will describe the recent Security Service Challenge (SSC) that enabled several EGI partners to investigate a simulated large-scale incident. After summarising the whole activity, we will present the essentials of digital forensics, focusing on collecting and analysing artefacts that could be observed during the SSC run.
      SSC Forensics Walkthrough
      We will describe the recent Security Service Challenge (SSC) that enabled several EGI partners to investigate a simulated large-scale incident. After summarising the whole activity, we will present the essentials of digital forensics, focusing on collecting and analysing artefacts that could be observed during the SSC run.

    • 09:00 13:00
      interTwin Tech F2F
    • 11:00 12:30
      Getting started with the EGI Workload Manager Service 1h 30m

      The DIRAC interware is a complete software solution for communities of users that need to exploit distributed, heterogeneous compute resources for big data processing.
      DIRAC forms a layer between the user community and the compute resources to allow optimized, transparent and reliable usage. DIRAC can connect various types of compute (Grids, Clouds, HPCs and Batch systems), storage, data catalogue resources. DIRAC is used by several of the most compute intensive research infrastructures. The EGI installation of DIRAC is used by diverse communities, such as WeNMR (Structural Biology), VIP (Medical imaging), Pierre Auger Observatory (astrophysics).
      In this session we will demonstrate the service functionalities with some basic examples running user applications and managing user data. The expected result of this training is that the users are able to start working with the EGI Workload Manager service.

      Speakers: Andrei Tsaregorodtsev (CNRS), Vanessa Hamar (CNRS)
    • 11:00 12:30
      Processing Data from EOSC on EGI Compute resources training 1h 30m

      Over the past decade, the scientific research had been revolutionized by the infrastructure of distributed computing and the concept of open data. Compute-intensive research has once been a barrier for many researchers, but today the cloud and distributed computing infrastructure can provide researchers immediate access to an unlimited amount of computing power and allows them to ask questions that may not have been possible. 

      Attendees will be guided through the necessary steps to find datasets in the EOSC Marketplace, gain access to EGI cloud resources to stage there the datasets they are interested in, and run scientific applications in the cloud to perform data transformation and analysis steps.

    • 11:00 12:30
      openEO Platform: Enabling analysis of large-scale Earth Observation data repositories 1h 30m

      openEO offers a unique opportunity to experience a hands-on training session of the cloud computing platform and API to access and analyse Earth Observation (EO) data. Through its client libraries in R, Python and Javascript, JupterLab environment and Web Editor, users can develop processing workflows interactively. Three federated backends provide a scalable analysis of EO data from pixel to continental scale, offering a unified and reproducible solution. D

      This training session will be an interactive and engaging experience, allowing participants to explore the openEO platform capabilities and its use cases. Through the session, participants are expected to gain hands-on experience in using the platform and its client libraries, developing processing workflows, and analysing large-scale EO data. The training session will also be a great opportunity for participants to ask questions and interact with the openEO team.