Open Source Initiatives

Community Releases

The products or outcomes of the R&D effort under NSM are made available to users at no cost in order to reach wider audiences as well as our contribution back to support both local and global HPC communities. The list of open-source solutions offered are listed below.

5

ParaS Compiler

ParaS Logo

This compiler is developed to facilitate device-agnostic programming based on SYCL 2020 specification. For installation, usage, and example codes, visit our GitHub Repository.

Download ParaS compiler from the GitHub repo above.

5

Optimized OpenBLAS

OpenBLAS Logo

This is an ARM-optimized version of OpenBLAS based on release 0.3.26, featuring ARM-specific vectorization enhancements, which are also included in the official OpenBLAS 0.3.29 release. For installation and usage visit our GitHub Repository.


Download the container from the GitHub repo above.

5

ASTViz

Tool for visulization of Clang AST using graphs on browser window. For installation and usage visit our GitHub Repository.







Download the container from the GitHub repo above.

Contact us at paras@cdac.in for any issues and feedback.

HPC Software Stack

Indigenous Software Stack Development:

As part of the NSM, a sophisticated HPC software stack  was indigenously developed utilizing open-source technologies. This suite was engineered to meet the diverse requirements of NSM systems, enhancing performance and ensuring seamless integration across HPC environments.

PARAM RUDRA Software Stack 2025

Core Components of the NSM HPC Software Stack are:
  • Operating System: CentOS/Alma has been selected as the foundational operating system, providing a stable and robust platform for HPC operations.
  • Drivers: The suite includes critical drivers such as CUDA for GPU acceleration, alongside network and storage drivers, to facilitate efficient data management and connectivity.
  • File Systems: Support for both local file systems and Lustre, a high-performance distributed file system, is integrated to manage extensive data volumes effectively.
  • Provisioning: The xCAT (Extreme Cluster Administration Toolkit) is employed for provisioning, automating the deployment and configuration of cluster resources.
  • Resource Management: SLURM (Simple Linux Utility for Resource Management) is utilized for job scheduling and resource management, optimizing the allocation and use of resources.
  • Cluster Management: It is achieved through tools such as Ganglia and Nagios, supplemented by C-DAC tools and Osticket for efficient issue tracking and support.
  • Communication Libraries: Essential communication libraries, including MPI (Message Passing Interface), Intel MPI, MVAPICH, and PGAS (Partitioned Global Address Space), are integrated to enable effective parallel communication and data exchange.
  • Development Tools: The suite features GNU Compiler & Intel oneAPI Compilers to support a broad spectrum of programming languages and optimizing computational performance.
  • Application Libraries: Key libraries such as NetCDF for network data, mathematical libraries for numerical computations, GNU libraries, and Deep Learning/Machine Learning (DL/ML) libraries are included to support a wide range of application needs.
  • Visualization Tools: Scientific visualization and data analysis are facilitated through tools like GNU Plot, ParaView, and VMD (Visual Molecular Dynamics), enabling detailed examination and interpretation of simulation results.
  • Performance Monitoring: The software stack incorporates performance monitoring tools and benchmarks to continuously evaluate and enhance system performance.
  • Software Ecosystem: The creation of software products like C-CHAKSHU, CHReME, HPC Tutor  PARADE, CAPC, User Creation Portal and Ticketing Support platforms reflects our emphasis on nurturing a robust software ecosystem. These solutions cater to diverse needs ranging from computational analysis and data visualization to project management, thereby augmenting research efficiency and productivity across scientific disciplines.
    • C-CHAKSHU: C-Chakshu is an HPC multi-cluster monitoring and Management platform that provides a unified dashboard over the web for all NSM sites with different geographic locations across India. Its web-enabled interface manages clusters of varied configurations and facilitates researchers and scientists of varied domains to run their applications with minimal efforts.
    • CHReME: To access Linux-based HPC environments, scientists and researchers require expertise in Linux and HPC, which is something many do not have. CHReME empowers users with an intuitive GUI to exploit HPC resources and provides a layer of abstraction to shield them from the complexity of accessing HPC resources. This enables them to concentrate on their core research/scientific work. It’s Web Interface makes clusters of different magnitudes easy to manage and monitor, which makes things easier for the HPC systems administrator while enabling researchers and scientists of varied domains to carry out their scientific simulations with minimal effort.
    • HPC Tutor: HPC Tutor is a revolutionary web based blended learning platform, designed to empower novice users with hands-on experience in the realm of High-Performance Computing (HPC). It seamlessly integrates theoretical knowledge with practical application, offering an immersive and educational journey in HPC Domain. This platform presents a unified web page, merging comprehensive presentation materials seamlessly with direct shell access to HPC clusters. This unique feature ensures a user-friendly and cohesive learning experience.
    • User Creation Portal : The User Creation Portal streamlines the account creation process by automating data collection and eliminating manual paperwork. It also enhances transparency by enabling users to monitor the progress of their requests and receive timely updates via email. This centralized platform ensures data accuracy and integrity, as users are required to provide official information verified through secure authentication mechanisms. Moreover, the portal offers a robust access control system, ensuring that sensitive user data is accessible only to authorized administrative personnel. This not only maintains data privacy but also facilitates efficient user management and oversight. By consolidating user information in a single location, the portal simplifies account auditing, reporting, and compliance processes, thereby improving overall operational efficiency and security.
    • PARADE: ParaDE is an environment for creating parallel applications on HPC platforms. Being web-based, it helps to access HPC irrespective of geographical location and carry out complete software development from compiling, debugging, automatic job submission, and project management, using multiple parallel paradigms such as OpenMP, MPI, etc. It enables users to just log in to the environment and use all software on the cluster without the hassles of installation or configuration on the client machine
    • CAPC: CAPC is an innovative solution that automatically parallelizes sequential code to different types of target parallel hardware without any inputs/contributions from the user. It liberates the application developer from the complexities of learning new parallel paradigms to write parallel programs for multicore and GPU parallel architectures
    • OS Ticket:  It is a ticketing system that streamlines HPC user support by routing inquiries, allowing for customization of data collection, and preventing duplicate responses from agents. It empowers agents with fast responses, SLA management, and collaboration tools. Users benefit from a quick response portal. This Desk provides valuable reporting and scales with the NSM vision, making it a powerful asset for managing user support. This HPC software stack ensures optimal performance and compatibility with NSM systems. It also aligns with India’s strategic goals of fostering technological self-reliance. By leveraging open-source technologies, the stack provides a flexible and cost-effective solution, driving innovation and bolstering capabilities within the national research and academic communities.

This HPC software stack ensures optimal performance and compatibility with NSM systems. It also aligns with India’s strategic goals of fostering technological self-reliance. By leveraging open-source technologies, the stack provides a flexible and cost-effective solution, driving innovation and bolstering capabilities within the national research and academic communities.

The deployment of this software suite has significantly enhanced the computational capabilities of NSM systems, empowering researchers and scientists to conduct high-impact research and achieve substantial advancements across various scientific and engineering disciplines.

This suit was deployed across all NSM HPC Facilities

Trinetra

Trinetra

Modern supercomputers are behind the scene powerhouses for scientific and human advancement applications such as weather prediction, Numerical simulations, bio-informatics, and material research. CDAC was entrusted with design and development of indigenous core technologies for supercomputing and has built several generations of PARAM supercomputing systems. One of the core components of a PARAM supercomputer is PARAMNet interconnect network. Under the NSM project, CDAC has developed three generations of latest PARAMNet interconnects called as Trinetra series of networks. Trinetra design and development effort spans multiple domains related with state of art hardware system design. It has multiple hardware and software components, such as NCC (Network Controller Chip: VLSI communication co-processor design), NIC (Network Interface Card), and LWP (Lightweight protocol) software. Together, these components realize a high bandwidth, low latency, scalable network fabric supporting industry standard programming interfaces.


Trinetra development has been done in three phases
  • Trinetra-POC: This was proof of concept phase, with focus being getting know-how about switchless networks, 3D Torus topology, and onload/hybrid architecture of Network Controller Chip. The hardware used was based 40Gbps physical link layer for IO and PCIe Gen-3.0 2x interface for host interface. The PoC hardware was used for extensive validation of key architectural concepts and hardware/software co-verification.
  • Trinetra-A is based on100Gbps physical link layer, PCIe Gen-3.0 x8 host interface, and uses 3D Torus as network topology. All hardware, software and firmware components of the network have been developed successfully. Trinetra-A is currently deployed in PARAM Rudra 1PF system at CDAC Pune.
  • Trinetra-B is the latest hardware developed with performance and topology enhancements, leveraging experience gained during PoC and TrA development phases. TrB is based around PCIe Gen-3.0 x16 host interface, and 200Gbps physical link layer. Ten links @ 200Gbps allow for Dragonfly/ Supercluster topology. TrB planned to be deployed in PARAM 20PF system.

Trinetra development effort is expected to continue under forthcoming NSM2.0 with primary aim of indigenous development of Exascale systems.

  • Use Cases
    • HPC and Enterprise applications
  • Salient Features
    • High Performance: High bandwidth and low latency, supported by innovative hardware capabilities and a lightweight software stack.
    • Scalability and Power Efficiency: Designed to ensure real performance, scalability, and power efficiency, while facilitating experimentation and application optimization.
    • Topology Flexibility: Supports 3D Torus (Trinetra-A) and Dragonfly/Supercluster (Trinetra-B) topologies for large-scale scalability eliminating need for dedicated switches.
    • Multi-Processor Support: Compatible with multiple processor and platform architectures.
    • Industry Standards Compliance: Supports standard HPC programming interfaces like MPI and legacy socket interfaces.

Technical Specifications

Trinetra-A:

  • Six 100Gbps full duplex interfaces
  • PCIe GEN-3.0 x8 host interface
  • 3D Torus topology
  • NCC-I Co-processor

Trinetra-B:

  • Ten 200Gbps full duplex interfaces
  • PCIe GEN-3.0 x16 host interface
  • Dragonfly/ Supercluster topology
  • NCC-II Co-processor

System Software
  • Includes device drivers and kernel/user components, with OFED support for seamless integration with existing systems.

Rudra-SPX

Gen2 (Rudra-SPX)

Rudra-SPX is based on Intel’s Eagle stream Platform. It is designed with Intel’s fourth generation Xeon Scalable Processor (SP) Sapphire Rapids (SPR) with Intel Chipset Emmitsburg. It supports next-gen industry-standard technologies such as PCIe 5.0, HBM & DDR5 memory speeds of up to 5600MT. It is compatible with Open Compute Platform (OCP) 48V DC for Cloud/Data center and HPC market. For cooling it has option of both Liquid (better energy efficiency) and Air cooling. The server is designed with interfaces required for HPC system and in small form factor.


Main features of Rudra-SPX
  • Fifth/Fourth Generation Xeon Emerald Rapids/ Sapphire Rapids Server Processor with 56 cores
  • Dual Sockets with TDP support up to 300W per CPU.
  • Memory: DDR5, RDIMM 4800 MT/s 1-SPC, 8 Channels per socket supporting up to 2 TB/socket.
  • Other Interfaces supported:
    • One OCP NIC 3.0 cards
    • Four PCIe Gen5.0 (32GT/s) x16 ports,
    • Three UPI 2.0 links support per CPU (12.8 GT/s, 14.4 GT/s, 16 GT/s)
    • Two (stacked) M.2 NVMe
    • One x8 PCIe 3.0 Ethernet card

This server is compliant with EMI/EMC and safety certifications such as CE, FCC, BIS and UL.

Its development and manufacturing has been instrumental in capability building, boost to manufacturing industry, Partner development in India.

Versatile design of this server is allowing possibility of many variants


System Management Firmware and BIOS

System Management Firmware based on Intel OpenBMC

Main Features:

  • ACPI power management.
  • Out-of-band hardware health monitoring.
  • Event logging
  • Power and thermal management.
  • iKVM / virtual media support.
  • Intel PFR and seamless update with cryptographically signed firmware.

Different product SKUs available are
  • 1OU servers (two servers in one chassis) with centralized Power supply
  • 5OU servers (two servers in one chassis) with centralized Power supply

This server is designed to support both air-cooling and liquid cooling server configurations

Rudra-GRX-AP

Gen3 (Rudra-GRX-AP)

Rudra-GRX-AP will be based on Intel’s Birch Stream Platform. It is designed with Intel’s sixth generation Xeon Scalable Processor (SP) Granite Rapids (GNR)-AP. It supports next-gen industry-standard technologies such as PCIe 5.0 & DDR5 memory speeds of up to 8800MT. It is compatible with Open Compute Platform (OCP) 48V DC for Cloud/Data center and HPC market. For cooling it has option of both Liquid (better energy efficiency) and Air cooling. The server is designed with interfaces required for HPC system and in small form factor.


Main features of Rudra-GRX-AP
  • Sixth Generation Xeon Granite Rapids Server Processor with up to 120 cores
  • Dual Sockets with TDP support up to 500W per CPU.
  • Memory support: DDR5; 12 Channels of DDR5 per socket supporting up to 3TB/socket. Memory Speed DDR5 RDIMM 6400 MT/s 1-SPC; 8800 MCR DIMM
  • DC-SCM 2.0 Specification compatible
  • Other Interfaces supported:
    • Two OCP NIC 3.0 cards
    • Up to 8 PCIe Gen5.0 (32GT/s) x16 ports,
    • Up to 4 x16 CXL 2.0 Ports
    • Up to 6 UPI 2.0 links support per CPU (2.5 GT/s, 16 GT/s, 20 GT/s, and 24 GT/s).

System Management

The BMC will be on DC-SCM V2.0 Card. ASpeed AST2600 Baseboard Management Controller (BMC) will be used for out-of-band system management. Major management functionalities include  

  • ACPI power management.
  • Out-of-band hardware health monitoring.
  • Power and thermal management.
  • iKVM / virtual media support.
  • Intel PFR and seamless update with cryptographically signed firmware.
  • IPMI 2.0 compliant.
  • Redfish® (RESTful API) support.
  • Web UI – Vue as web interface.
  • Remote debug – Intel ASD and ACD.
  • Intel Node Manager support.
  • Compliant with standard IPMI client tools like ipmitool.
  • Dedicated management Ethernet with NSCI fail-over management network.

Form Factor
  • Rudra-GRX-AP-2110LB 1OU Liquid cooled: Only CPUs
  • Rudra-GRX-AP-2140AB 4OU Air cooled: CPU-GPU
  • Rudra-GRX-AP-2140LB 4OU Liquid cooled: CPU-GPU

Rudra-I

Gen1 (Rudra-I)

Under the build approach C-DAC has designed and developed a compute server “Rudra” as one of the most important subassemblies required for Supercomputers. Rudra, C-DAC’s indigenous server platform features Intel Xeon 2nd Generation Cascade Lake processor supporting DDR4 memory, with two expansion slots for GPU and Accelerator cards. C-DAC is in process of deploying HPC Systems across the nation with compute performance totalling 40PF, based on Rudra Servers called PARAM Rudra.

C-DAC has partnered with EMS partners in India for mass manufacturing of Rudra servers. More than 1300 servers have been produced so far and are getting deployed at different installation sites. Under transfer of technology agreement, C-DAC has also transferred the Rudra Server Design to three industry partners, viz. M/s VVDN, M/s Kaynes technologies and M/s Avalon technologies. All these activities have catalysed the development of entire eco-system right from design, manufacturing and userbase generation.


Main features of Rudra-I
  • Suitable for diverse applications such as, High performance computing, cloud and edge computing and communication
  • Dual socket, 2nd Gen Intel Xeon Scalable Processors (upto TDP 165W)
  • 20 DDR4 DIMM slots LRDIMM, RDIMM (total upto 1.28TB)
  • Support for Intel Optane Persistent Memory (NVDIMM)
  • Two PCIe 3.0 x16 slots with GPU/Accelerator support
  • Onboard 2 x 10Gbps ethernet Interfaces
  • One additional high speed network slot (Upto 100Gbps, OCP NIC 3.0 compatible)
  • One 1GbE BMC management port
  • Support for IPMI2.0
  • Support for KVM
  • Support for two U.2 NVMe/SATA

This server is compliant with EMI/EMC and safety certifications such as CE, FCC, BIS and UL.

Its development and manufacturing has been instrumental in capability building, boost to manufacturing industry, Partner development in India.

Versatile design of this server is allowing possibility of many variants


Different product SKUs available are
  • 2U server with Built-in power supply (can go as standard 19” server supporting 2 GPUs)
  • 2U storage server with Built-in power supply
  • 1U servers (two servers in one chassis) with centralized Power supply
  • 2U servers (two servers in one chassis) with centralized Power supply
  • ½ width 1U with centralized Power supply
  • ½ width 2U with centralized Power supply

This server is designed to support both air-cooling and liquid cooling server configurations

Products

Products

The National Supercomputing Mission (NSM) initiated by the Government of India aims towards achieving the goals of attaining self-reliance in supercomputing and problem solving in various domains of scientific and technological endeavors.

Under this mission, C-DAC is creating an infrastructure of family of supercomputers of varying capacities spanning a wide range. One of the key modalities in this mission is to use indigenous build approach, wherein design and manufacture of substantial sub-assemblies would be carried out in India.

Gen1 (Rudra-I)
5

Rudra-I

Under the build approach C-DAC has designed and developed a compute server “Rudra” as one of the most important subassemblies required for Supercomputers. C-DAC is in process of deploying HPC Systems across the nation with compute performance totalling 40PF, based on Rudra Servers called PARAM Rudra.
5

Rudra-SPX

Rudra-SPX is based on Intel’s Eagle stream Platform. It supports next-gen industry-standard technologies such as PCIe 5.0, HBM & DDR5 memory speeds of up to 5600MT. It is compatible with Open Compute Platform (OCP) 48V DC for Cloud/Data center and HPC market.
5

Rudra-GRX-AP

Rudra-GRX-AP will be based on Intel’s Birch Stream Platform. It supports next-gen industry-standard technologies such as PCIe 5.0 & DDR5 memory speeds of up to 8800MT. It is compatible with Open Compute Platform (OCP) 48V DC for Cloud/Data center and HPC market.

Trinetra HPC network is latest in PARAMNet series of high bandwidth, low latency Indigenous HPC network developed by C-DAC. It consists of hardware and software components working seamlessly to realize large scale PARAM supercomputing systems. The network is transparent to end uses, offering industry standard API interfaces to run HPC and Enterprise applications without any end-user porting effort.

Trinetra-A
5

Trinetra-A

It is based on 100Gbps physical link layer, PCI-e Gen3 X8 host interface, and uses 3D Torus as network topology. All hardware, software and firmware components of the network have been developed successfully. It is currently deployed in PARAM Rudra 1PF system at CDAC Pune.
5

Trinetra-B

It is the latest hardware developed with performance and topology enhancements, leveraging experience gained during PoC and TrA development phases. TrB is based around PCI-e Gen3, x16 host interface, and 200Gbps physical link layer. Ten links @ 200Gbps allow for Dragonfly/ Supercluster topology. TrB planned to be deployed in PARAM 20PF system.

As part of the NSM, a sophisticated HPC software stack  was indigenously developed utilizing open-source technologies. This suite was engineered to meet the diverse requirements of NSM systems, enhancing performance and ensuring seamless integration across HPC environments.

5

Indigenous Software Stack Development:

The deployment of this software suite has significantly enhanced the computational capabilities of NSM systems, empowering researchers and scientists to conduct high-impact research and achieve substantial advancements across various scientific and engineering disciplines. This suit was deployed across all NSM HPC Facilities

Research & Development

Research and Development in Next Generation HPC Technologies

Research and Development is being done for the next generation HPC technologies. Some of the areas include:

  • Developments for building state-of-the-art HPC machines
  • Exploring next generation Data centre technologies
  • Next generation system architectures and prototypes
  • Open Source Initiatives
  • Power optimization technologies
  • Programming paradigm for Exascale
  • R&D for high speed network scalable for Exascale supercomputers
  • Scalable Algorithm developments
  • Silicon photonics
  • Special purpose machines for Bio-informatics