High Performance Computing Facility

Master Node

02Nos x Intel Xeon Gold 6326 2.9G, 16C/32T, 11.2GT/s and 128 GB Memory, 4 x 2.4TB HDD, 12 Gbps RAID Controller (RAID 0,1,5,6,1+0) with 8 GB Cache, 1 x Dual Port 100GbE QSFP Adapter with QSFP28 SR4 100GbE Optic 2 x 1 Gbps network card Make & Model No – Dell PowerEdge R750 Server Serial No. 3R6LW14 01

CPU Compute Node

02Nos x Intel Xeon Gold 5318 2.1G, 24C/48T, 11.2GT/s, 256 GB memory, 2 x 960GB SSD Disk capacity, 12 Gbps RAID Controller (RAID 0,1,5,6,1+0) with 8 GB Cache, 1 x Dual Port 100GbE QSFP Adapter with QSFP28 SR4 100GbE Optic 2x 1 Gbps network card Make & Model No – Dell PowerEdge R650 Server Serial No.: ( 19JKW14, 29JKW14, 39JKW14, 49JKW14, 59JKW14, 69JKW14, F8JKW14 G8JKW14, H8JKW14, J8JKW14)

GPU Node

02Nos x Intel Xeon 5318 2.1G, 24C/48T, 11.2GT/s. 512 GB memory, 2 x 240 GB SSD for OS and 5 x 2.4TB for data, 12 Gbps RAID Controller (RAID 0,1,5,6,1+0) with 8 GB Cache, 1 x Dual Port 100GbE QSFP Adapter with QSFP28 SR4 100GbE Optic, 2×1 Gbps network card, 2 nos NVIDIA L40, PCIe, 300W, 48GB Passive, Double Wide, Full Height Make & Model No – Dell PowerEdge R750 Server Serial No.: ( 2R6LW14 ) 01

I/O Node

02Nos x Intel Gold 6326 2.9G, 16C/32T, 11.2GT/s, 128 GB Memory, 2 x 2.4 TB SSD for OS, 12 Gbps RAID Controller (RAID 0,1,5,6,1+0) with 8 GB Cache, HBA Adapter: Dual Port 32Gb Fibre Channel HBA, 1 x Dual Port 100GbE QSFP Adapter with QSFP28 SR4 100GbE Optic, 2x 1 Gbps network card Make & Model – Dell PowerEdge R740 Server Serial No.: ( 1R6LW14, JQ6LW14 )

PFS Storage

PFS Storage: – Controller Cards 32Gb FC Type-B 8 Port Dual Controller 1 16G/SFP and 32G/SFP+ FC Optics 2X SFP+, FC32, 32GB 04Nos Multi-Mode FC LC-LC Cables Dell Networking Cable, OM4 LC/LC Fiber Cable, (Optics required), 2 Meter 8Nos Hard Drives 20TB HDD SAS ISE 12Gbps 7.2K 512e 3.5in Hot-Plug, AG Drive 8Nos Hard Drives 1.92TB SSD SAS ISE, Read Intensive, up to 24Gbps 512e 2.5in with 3.5in HYB CARR, AG Drive 4Nos Power Supply Power Supply, 580W, Redundant, WW 1 Make & Model – Dell ME5012 Storage Array Serial No.: ( BV2C324 )

Primary Interconnect

Dell EMC S5232F-ON Switch, 32x 100GbE QSFP28 ports, IO to PSU air, 2x PSU 1 Ethernet Optics Dell Networking, Transceiver, 100GbE QSFP28 SR4 Make & Model – Dell EMC S5232F-ON Switch Serial No.: ( 7X56SR3 ) 01

Management Switch

24-Port Gigabit Ethernet Smart Switch with 2 SFP Ports. Make & Model – NETGEAR -GS724T Switch SR No: 5XF6385CA01B1

KVM Switch with the Console Cable

Make : ATEN Model : Aten-CS1316 16-port Combo KVM S/N:-Z3NA-115-0178

KVM Display

Make: ATEN Model : Aten-CL1000N 19″ LCD KVM Console, 19″ LCD Panel, keyboard, touchpad, 1U rack mount form factor S/N:-Z8N7-052-0009

Smart rack with Integrated cooling

3 x 7KW Cooling unit 2 x 20 KVA UPS 2 Racks Solution Make& Model : VERTIV Smart Rack SCP Serial No – 202403SCP0005

Application Form to Apply for HPCF User ID

RACHEL HPC FACILITY Resource Usage Policy

Policy Overview

Date: 11 March 2026

The Rachel High Performance Computing (HPC) Facility provides shared computational infrastructure to support the research, teaching, and scholarly activities of the institution. To ensure equitable access and the efficient utilization of these resources, all registered users are required to read, understand, and comply with the provisions set out in this Policy. These guidelines are effective immediately upon account activation and apply to all submitted workloads, stored data, and system interactions.

ParameterLimit / Requirement
Maximum Job Runtime120 hours (5 consecutive days)
Per-User Storage Quota150 GB (unauthorised use prohibited)
Job MonitoringUser's ongoing responsibility
Storage MaintenancePeriodic review & cleanup required
Non-Compliance ActionJob termination / access suspension

Job Runtime Limit

No user shall be permitted to execute any computational job on the Rachel HPC cluster for a continuous period exceeding total 120 hours (CPU-5 consecutive days or GPU-3 consecutive days). Jobs that approach or exceed this threshold will be subject to administrative intervention, including automatic or manual termination, to preserve system availability for all users.

Users requiring extended runtime beyond the standard limit must submit a written request to the System Administrator prior to job submission, providing scientific justification and an estimated resource budget.

Storage Utilization Limit

Individual users shall not utilize more than total 150 GB (for CPU nCPU-12 Core & for GPU nGPU-6 Core) of storage on the Rachel HPC system without prior written authorization from the System Administration team. This quota encompasses all personal directories, project workspace, and temporary scratch areas associated with the user account.

The following storage hygiene practices are mandatory:

  • Removal of completed job output files that are no longer required for active analysis.
  • Deletion of temporary, intermediate, and scratch files upon job completion.
  • Regular archival of results to institutional long-term storage systems.
  • Prompt response to storage-usage alerts issued by System Administration.

Fair Resource Usage

All users are expected to exercise good judgement in their use of Rachel HPC resources. Users shall not monopolies CPUs, memory, network bandwidth, or storage in a manner that disrupts or prevents other researchers, faculty members, or students from accessing the facility.

Submitting large numbers of simultaneous jobs that saturate the scheduler queue without scientific necessity, or deliberately circumventing resource allocation controls, constitutes a violation of this Policy.

Job Monitoring Responsibility

Users bear full responsibility for monitoring all jobs they submit to the Rachel HPC scheduler. This responsibility includes:

  • Regularly checking job status, output logs, and resource-consumption metrics.
  • Promptly terminating any job found to be stalled, malfunctioning, or producing incorrect results.
  • Ensuring that submitted jobs do not run beyond their stated resource requirements.

Users who are unable to actively monitor a job due to absence or other commitments should make arrangements for a designated colleague to assume monitoring responsibility, or should limit job scope accordingly.

Storage Maintenance

Users shall periodically review their allocated storage space and proactively remove unnecessary files, redundant datasets, temporary data, and intermediate outputs. Maintaining adequate free capacity on the Rachel HPC storage systems is a shared responsibility that directly impacts the productivity of all users.

System Administration may issue storage-usage warnings when a user's allocation approaches the permitted quota. Users are expected to respond to such warnings within five (5) working days by reducing their storage footprint to within the authorized limit.

Policy Compliance

All users of the Rachel HPC facility are required to comply with the provisions of this Policy as a condition of continued access. Compliance is the personal responsibility of every account holder.

âš   Non-Compliance & Administrative Authority Failure to comply with any provision of this Policy may result in the immediate termination of running jobs, restriction of computational resource allocations, or temporary suspension of access privileges to the Rachel HPC facility. The System Administrator reserves the right to terminate any user process deemed to be making undue or excessive use of system resources, without prior notice, in order to maintain overall system stability and equitable access for all users.

                                                                                                                              

Signature of HOD (CSE)