Madhava HPC Cluster
Master/Login/Service nodes: 1
CPU nodes: 31
GPU accelerated nodes: 2
Storage nodes: 2
Madhava HPC cluster is an aggregation of a large number of computers connected through networks. The basic purpose of the master node is to manage and monitor each of the constituent component of Madhava HPC from a system’s perspective. This involves operations like monitoring the health of the components, the load on the components, the utilization of various sub-components of the computers in Madhava HPC. This also works as a login node which used for administrative tasks such as editing, writing scripts, transferring files, managing your jobs and the like. You will always get connected to one of the login nodes. From the login nodes you can get connected to a compute node and execute and interactive job or submit batch jobs through the batch system (SLURM) to run your jobs on compute nodes. For ALL users Madhava HPC Cluster login nodes are the entry points and hence are shared. This node also works as a service node to provide Job Scheduling Services and other services to the cluster
Master/ Login/service Nodes : 1 |
|
---|---|
Model |
Lenovo ThinkSystem SR650 |
Processor |
2 nos of Intel(R) Xeon(R) Gold 6248 CPU @ 2.50GHz. |
Cores |
64 cores |
RAM |
192 GB |
HDD |
2 TB |
CPU nodes are indeed the work horses of Madhava HPC. All the CPU intensive activities are carried on these nodes. Users can access these nodes from the login node to run interactive or batch jobs. by users in the aforementioned way.
CPU Only Nodes : 31 |
|
---|---|
Model |
Lenovo ThinkSystem SR630 |
Processor |
2 nos of Intel(R) Xeon(R) Gold 6248 CPU @ 2.50GHz. |
Cores |
40 cores |
RAM |
192 GB |
HDD |
1 TB |
GPU compute nodes are the nodes that have CPU cores along with accelerators cards. For some applications GPUs get markedly high performance. For exploiting these, one has to make use of special libraries which map computations on the Graphical Processing Units (Typically one has to make use of CUDA or OpencCL).
CPU Only Nodes : 31 |
|
---|---|
Model |
Lenovo ThinkSystem SR650 |
Processor |
2 nos of Intel(R) Xeon(R) Gold 6248 CPU @ 2.50GHz. |
Cores |
40 cores |
RAM |
192 GB |
HDD |
1 TB |
GPU |
2* Tesla V100-PCIE-32GB per node |
GPU cores per node |
2* 5120 |
GPU Tensor Cores |
2* 640 |
Providing the storage service to head and compute nodes using GPFS
Storage/ IO nodes:2 |
|
---|---|
Model |
Lenovo ThinkSystem SR630 |
Processor |
2 nos of Intel(R) Xeon(R) Gold 6248 CPU @ 2.50GHz. |
Cores |
40 cores |
RAM |
192 GB |
HDD |
1 TB |
It's a backend to the IO nodes which is providing storage service to head and compute nodes
Storage/ IO nodes:2 |
|
---|---|
Model |
ThinkSystem DE6000H 2 nos of DE Controller with 16 GB, 60 Chasis. |
Home Space |
204 TB |
HDD |
Scratch Space 60 TB |
Address:
National Institute of Technology Calicut,
NIT Campus P.O 673 601,
Kozhikode, India.
Departments and Centres
Committees / Cells
Achievements
Minutes & Reports
Sponsored by NITCAA - Alumni of B.Tech 1993-97 Batch.
Designed and Developed by Xpertconsortium Technologies LLP.
Maintained by C-PRIME and CITRA, NITC.