Roadmap

Now

  • High-Level Training Program Development

    We will continuously develop user training activities to ensure that end users have the knowledge and skills required to effectively use Sigma2 national services. This will contribute to develop national AI skills and plan AI development campus. This includes preparing instructional materials, conducting sessions, and gathering user feedback for continuous improvement.

  • Access to Quantum Computing

    Sigma2 will provide Norwegian researchers with easy access to EuroHPC quantum computers, including LUMI-Q, as well as local quantum emulators for testing and learning. Access will be available through the same interfaces you already use (Waldur and Sigma2 cloud services), with clear documentation, examples, and training materials. You will be able to request access to quantum resources directly through Sigma2 service portal, run quantum algorithms on LUMI-Q or experiment with them first on local emulators, combine classical HPC and AI workflows with quantum steps in a seamless way, and make use of educational notebooks, small demos, and ready-to-use templates that help you get started quickly. This work ensures that quantum computing becomes a natural extension of Sigma2’s services easy to access, easy to learn, and fully integrated into your existing research workflow.

  • Introduce interactive compute partition on HPC

    We are restructuring how we handle light-weight, interactive workloads. By moving development and light-weight testing tasks off login nodes and into a dedicated, (possibly quota-free) partition, we ensure system stability while providing users with a responsive environment for their day-to-day computing needs.

  • Support the Data Rescue initiative

    We will assist the Norwegian research community in safeguarding datasets currently stored abroad, which may be at risk of deletion or loss.

  • Introducing heatmaps for NIRD projects in MAS

    To improve visibility and decision-making at the project level, we will introduce heatmaps in MAS that visually represent storage usage patterns. This feature will empower project leaders to gain a deeper understanding, effectively manage, and optimize data storage resources.

  • GPU support in our administrative system (MAS)

    We are upgrading our administrative system (MAS) to recognize GPU resources as a distinct resource type alongside CPU. This initiative ensures that GPU consumption - whether on national machines or external systems like LUMI - is allocated, tracked, and reported with the same visibility as traditional CPU resources.

  • Integrate our administrative system (MAS) with Puhuri

    Puhuri is the name of the portal used to administer LUMI projects. We are working on integrating this portal more tightly with our existing administrative system (MAS). This will ensure a more seamless experience for Sigma2 users when utilizing LUMI and the national resources.

  • Implement data decommissioning policy

    While a formal data decommissioning policy is in place, consistent execution and follow-up have been lacking. A revision is necessary to incorporate and enable automated maintenance procedures, which will ensure reliable adherence to retention schedules.

  • NIRD Research Data Archive development

    Use cases and functionalities which were not handled as part of the baseline product development during the Archive2021 project will be addressed as part of the NIRD Research data Archive product development process. API development will be prioritized to allow metadata harvesting by metadata registries. Functionalities such as customized data ingestion, customized search and reporting will be addressed.

  • Enable S3 access from HPC systems

  • Allow the Service Platform to submit jobs on the HPC clusters

    The scientific workflows are becoming more complicated. Users produce and analyze large data volumes on the HPC system, store the data on NIRD and post-process the data on the NIRD Service Platform. There is the need for a continuum of analytical environments at the scale. Allowing job submission from the NIRD Service Platform to the HPC would enable scale out of the workflows.

  • Enhancing the FAIRness of the NIRD Research Data Archive

    We will improve the interoperability and findability of the NIRD Research Data Archive to enhance its overall FAIRness, making research data easier to discover, access, and integrate across systems. The Research data Archive will continue to evolve with new features and stronger support for FAIR principles. Development will be co-designed with researchers and key stakeholders to ensure it meets community needs.

  • Enhance integration between NIRD and the HPC clusters

    Data cannot and should not be permanently stored on the HPC clusters. We will implement a fast, reliable, and intuitive data staging solution to simplify and accelerate data movement between the HPC clusters and NIRD mid and long-term storage.

  • Data-centric model strategies

    To fully utilize advancements in infrastructure and new system capabilities, the data-centric model must be consistently maintained and periodically updated. Regular updates ensure that the model stays responsive to evolving technological features, allowing users to harness the potential of new functionalities. This proactive approach optimizes performance and keeps the model aligned with the latest operational improvements. The new model shall take into consideration the current and evolving service landscape from Tier-2 (at the local universities) to the Tier-1 (Betzy, Saga, Olivia) and to the Tier-0 (LUMI). A look at the role of the cloud storage in picture is also enviable.

  • Open, confidential and restricted data services

    The data landscape is becoming more complex with the explosion of new technologies and research frontiers. AI/ML is just paradigmatic of this revolution. A variety of data with different requirements with regard to access and protection is produced. We will design and adapt the services to data which require confidentiality while still be free from regulations related to health personal sensitivity.

Next

  • Enable storage solution for large language models

    We will investigate and implement solutions to support a Git-like filesystem and integrate it with the NIRD Central Data Library service to support large language model repositories.

  • Streamlined S3 account management

    We will simplify and centralize the S3 account management processes on NIRD. This strategic change will strengthen security protocols, establish uniform control, and drive down both administrative and operational overhead.

  • Streamlined quota management

    We will simplify and automate storage quota management, improving transparency and ensuring better control over resource usage on NIRD services (For eg: Data Peak, Data Lake).

  • Achieving the ISO 27001 certification

    We will strengthen our information security practices and build customer trust by obtaining ISO 27001 certification, an internationally recognized standard for Information Security Management Systems.

  • Introduce a common contribution model for all storage services

    We need a common storage contribution model covering all our services offered, including HPC storage. The contribution model will be aligned with the data-centric model.

  • Research Data Archive for confidential data

    The paradigm is to make Norwegian scientific data as open as possible, and as restricted as necessary. Finding mechanisms to archive, publish and share large volumes of unstructured scientific data with some requirements for higher security and access controls is still a challenge. Through domain expert and national provider consultation we will investigate and develop a comprehensive architectural design which can support both open research data and facilitate controlled access to sensitive and restricted research assets, ensuring privacy and compliance are baked into the design.

  • Provide a quantum computing service

    It is planned that Sigma2 will get access to quantum computing resources in 2026 through the LUMI consortium (VLQ). This initiative concerns facilitating access to these resources.

  • More automatic administration of HPC quotas

    We are implementing a set of automated administrative protocols in our administrative system (MAS) to improve the user experience. By shifting routine tasks such as account management and quota monitoring to automated systems, we ensure that the users are actively notified of relevant events. In addition, resources remain accessible to active projects while maintaining strict security and efficiency standards.

  • Implement Data Insight service

    We will focus on implementing solutions that offer comprehensive insights to service owners, operators and end-users of the Sigma2 infrastructure. We aim on generating periodical and on-demand reports on usage patterns, metadata, data type and much more.

  • Introduction of Research Objects

    We will address data management needs arising from steadily increasing data volumes in close collaboration with user communities and key stakeholders. We will focus on adopting Research Objects (ROs) as a foundation for improved, future-oriented data management. ROs will first be introduced on NIRD for selected pilot projects and later expanded to all projects.

Later

  • LUMI-AI Federation with Sigma2 Cloud

    Description Integrate Sigma2’s cloud stack (Waldur, OOD, SP/OpenStack) with LUMI-AI to allow Norwegian users to request, manage, and run workloads on LUMI-AI tenants directly from Sigma2’s national cloud portal. The federation covers AAI, project mapping, accounting (MAS), job orchestration, and secure data paths, using LUMI-AI’s multi-tenant architecture and service-cluster capabilities. Need Origin * Users need access to AI-scale GPUs beyond national capacity. * EuroHPC/NAIF coordination requires federated access to shared AI infrastructure. * LUMI-AI exposes APIs suited for cloud portal integration, not just HPC logins. Why Priority * Time: Aligns with SP migration & Waldur maturity (2026 pilots). * Cost: Avoids duplicating GPU investments domestically. * Quality: Provides a unified, policy-compliant user experience with hybrid workflows (Cloud <--> HPC).

  • Increase NIRD performance

    The performance of the NIRD must closely align with the performance requirements imposed by the Olivia machine. Additionally, there is an anticipated increase in demand due to the scaling up of cloud resources.

  • Implementing backup for NIRD Data Lake

    We will establish an on-demand backup service for a limited number of NIRD Data Lake tenants to strengthen data resilience and business continuity, ensuring critical information can be quickly recovered in case of loss, corruption, or system failure.

  • Introduction of data management solution for the Central Data Library

    We will enhance the NIRD Central Data Library (CDL) service by integrating a robust data management solution that improves data accessibility, consistency, and lifecycle governance. This effort will extend the service’s capabilities, streamline data operations, and provide a foundation for future scalability and advanced analytics.

  • Implementing data deduplication

    We will increase the effort in identifying and work on removing duplicated data to improve storage efficiency, performance, and data quality across NIRD and other storage systems.

  • Introduce interactive HPC service

    This is a further development of the interactive compute partition, allowing subscriptions to the partition at the organizational level, e.g. via Feide.

  • Sustainable governance

    We will review and propose adjustments to the sustainability, support model, and governance of the new Research Data Archive. The aim is to explore opportunities to better utilize national competencies and resources, with the potential to contribute to a more robust and efficient long-term service.

  • Discuss collaborations and synergies with DataverseNO

    We will establish a framework for tighter collaboration with DataVerseNO, which is a popular platform among researchers and curators for archiving small datasets. Users will be guided and supported in scaling out from one service to the other.