Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revisionPrevious revision
Next revision
Previous revision
Last revisionBoth sides next revision
en:centro:servizos:servidores_de_computacion_gpgpu [2020/06/03 13:44] – [Service description] fernando.guillenen:centro:servizos:servidores_de_computacion_gpgpu [2023/10/11 13:56] – [How to connect the servers] fernando.guillen
Line 2: Line 2:
  
 ===== Service description ===== ===== Service description =====
- +==== Servers with free access GPUs ====
-Servers with graphic cards: +
- +
- +
-  * ''ctgpgpu2'': +
-    * Dell Precision R5400 +
-    * 2 x [[http://ark.intel.com/products/33082/|Intel Xeon E5440]] +
-    * 8 GB RAM (4 x DDR2 FB-DIMM 667 MHz) +
-    * 1 Nvidia GK104 [Geforce GTX 680] +
-    * Ubuntu 18.04 operative system +
-      * Slurm (//mandatory to queue jobs!//) +
-      * CUDA 9.2 (//Nvidia official repo//) +
-      * Docker-ce 18.06 (//Docker official repo//) +
-      * Nvidia-docker 2.0.3 (//Nvidia official repo//) +
-      * Nvidia cuDNN v7.2.1 for CUDA 9.2 +
-      * Intel Parallel Studio Professional for C++ 2015 (//single license! coordinate with other users!//) +
-  * ''ctgpgpu3'': +
-    * PowerEdge R720 +
-    * 1 x [[http://ark.intel.com/products/64588|Intel Xeon E52609]] +
-    * 16 GB RAM (1 DDR3 DIMM  1600MHz) +
-    * Connected to a graphical card extensión box with: +
-      * Gigabyte GeForce GTX Titan 6GB (2014) +
-      * Nvidia Titan X Pascal 12GB (2016) +
-    * Ubuntu 18.04 operative system +
-      * Slurm (//mandatory to queue jobs!//) +
-      * CUDA 9.2 (//Nvidia official repo//) +
-      * Docker-ce 18.06 (//Docker official repo//) +
-      * Nvidia-docker 2.0.3 (//Nvidia official repo//) +
-      * Nvidia cuDNN v7.2.1 for CUDA 9.2 +
-      * Intel Parallel Studio Professional for C++ 2015 (//single license! coordinate with other users!//) +
-      * ROS Melodic Morenia (//repositorio oficial de ROS//)+
   * ''ctgpgpu4'':   * ''ctgpgpu4'':
       * PowerEdge R730       * PowerEdge R730
Line 38: Line 8:
       * 128 GB RAM (4 DDR4 DIMM  2400MHz)       * 128 GB RAM (4 DDR4 DIMM  2400MHz)
       * 2 x Nvidia GP102GL 24GB [Tesla P40]       * 2 x Nvidia GP102GL 24GB [Tesla P40]
-      * Centos 7.4 +      * AlmaLinux 9.1 
-          * Docker 17.09 and nvidia-docker 1.0.1 +          * Cuda 12.0 
-          * OpenCV 2.4.5 +          * **Mandatory use of Slurm queue manager**
-          * Dliv, Caffe, Caffe2 and pycaffe + 
-          Python 3.4cython, easydict, sonnet +  HPC cluster servers[[ en:centro:servizos:hpc | HPC cluster ]] 
-          TensorFlow +  CESGA servers: [[ en:centro:servizos:cesga | Access procedure info ]]  
-  * ''ctgpgpu5'':+ 
 +==== Restricted access GPU servers  ==== 
 + * ''ctgpgpu5'':
       * PowerEdge R730       * PowerEdge R730
       * 2 x  [[https://ark.intel.com/products/92980/Intel-Xeon-Processor-E5-2623-v4-10M-Cache-2_60-GHz|Intel Xeon E52623v4]]       * 2 x  [[https://ark.intel.com/products/92980/Intel-Xeon-Processor-E5-2623-v4-10M-Cache-2_60-GHz|Intel Xeon E52623v4]]
       * 128 GB RAM (4 DDR4 DIMM  2400MHz)       * 128 GB RAM (4 DDR4 DIMM  2400MHz)
       * 2 x Nvidia GP102GL 24GB [Tesla P40]       * 2 x Nvidia GP102GL 24GB [Tesla P40]
-      * Ubuntu 16.04+      * Ubuntu 18.04
           * **Slurm as a mandatory use queue manager**.           * **Slurm as a mandatory use queue manager**.
           * ** Modules for library version management **.           * ** Modules for library version management **.
-          * CUDA 9.0+          * CUDA 11.0
           * OpenCV 2.4 and 3.4           * OpenCV 2.4 and 3.4
           * Atlas 3.10.3           * Atlas 3.10.3
Line 64: Line 36:
       * 192 GB RAM memory(12 DDR4 DIMM 2933MHz)        * 192 GB RAM memory(12 DDR4 DIMM 2933MHz) 
       * Nvidia Quadro P6000 24GB (2018)       * Nvidia Quadro P6000 24GB (2018)
-      * Nvidia Quadro RTX8000 48GB (2020)+      * Nvidia Quadro RTX8000 48GB (2019)
       * Operating system Centos 7.7       * Operating system Centos 7.7
           * Nvidia Driver 418.87.00 for CUDA 10.1           * Nvidia Driver 418.87.00 for CUDA 10.1
           * Docker 19.03           * Docker 19.03
           * [[https://github.com/NVIDIA/nvidia-docker | Nvidia-docker  ]]           * [[https://github.com/NVIDIA/nvidia-docker | Nvidia-docker  ]]
-  * ''ctgpgpu7'':  +  * ''ctgpgpu9'': 
-      * Server Dell PowerEdge R740 +      * Dell PowerEdge R750 
-      * 2 processors[[https://ark.intel.com/content/www/us/en/ark/products/193388/intel-xeon-gold-5220-processor-24-75m-cache-2-20-ghz.html|Intel Xeon Gold 5220]] +      * 2 [[ https://ark.intel.com/content/www/es/es/ark/products/215274/intel-xeon-gold-6326-processor-24m-cache-2-90-ghz.html |Intel Xeon Gold 6326 ]] 
-      * 192 GB RAM (12 DDR4 DIMM a 2667MHz)  +      * 128 GB RAM  
-      * 2 x Nvidia Tesla V100S 32GB (2019) +      * 2 x NVIDIA Ampere A100 80 GB 
-      * Operating system Centos 8.1 +      * AlmaLinux 8.6 
-          * **Slurm as a mandatory use queue manager**. +           NVIDIA 515.48.07 driver and CUDA 11.7 
-          * ** Modules for library version management **. +  * ''ctgpgpu10'': 
-          * Nvidia Driver 440.64.00 for CUDA 10.+      * PowerEdge R750 
-          * Docker 19.03 +      * 2 [[ https://ark.intel.com/content/www/es/es/ark/products/215272/intel-xeon-gold-5317-processor-18m-cache-3-00-ghz.html |Intel Xeon Gold 5317 ]] 
-          * [[  https://github.com/NVIDIA/nvidia-docker | Nvidia-docker  ]] +      * 128 GB  RAM  
-  * ''ctgpgpu8'':  +      * NVIDIA Ampere A100 80 GB 
-      * Dell PowerEdge R740 +      * Sistema operativo AlmaLinux 8.7 
-      * 2 processors  [[https://ark.intel.com/content/www/us/en/ark/products/193388/intel-xeon-gold-5220-processor-24-75m-cache-2-20-ghz.html|Intel Xeon Gold 5220]] +           Driver NVIDIA 525.60.13 and CUDA 12.0 
-      * 192 GB RAM (12 DDR4 DIMM a 2667MHz)  +  ''ctgpgpu11'': 
-      * 2 x Nvidia Tesla V100S 32GB (2019) +      Server Gybabyte  G482-Z54 
-      * Operating System Centos 8.1 +      2 x [[ https://www.amd.com/es/products/cpu/amd-epyc-7413 | AMD EPYC 7413 @2,65 GHz 24c ]] 
-          * **Slurm as a mandatory use queue manager**. +      256 GB RAM 
-          * ** Modules for library version management **+      4 x NVIDIA Ampere A100 de 80 GB   
-          Nvidia Driver  440.64.00 for CUDA 10.2 +      AlmaLinux 9.1 
-          Docker 19.03 +           * Driver NVIDIA 520.61.05 and CUDA 11.8 
-          * [[  https://github.com/NVIDIA/nvidia-docker Nvidia-docker  ]]+  ''ctgpgpu12'': 
 +      Servidor Dell PowerEdge R760 
 +      * 2 procesadores [[ https://ark.intel.com/content/www/xl/es/ark/products/232376.html |Intel Xeon Silver 4410Y ]] 
 +      * 384 GB de memoria RAM  
 +      * 2 x NVIDIA Hopper H100 de 80 GB 
 +      * Sistema operativo AlmaLinux 9.2 
 +           * Driver NVIDIA 535.104.12 para CUDA 12.2 
 ===== Activation ===== ===== Activation =====
-All CITIUS users can access this service, but as not all servers are available all the time you have to register beforehand filling the [[https://citius.usc.es/dashboard/enviar-incidencia| requests and problem reporting form]]. +Not all servers are available to use freely. Access must be requested filling the [[https://citius.usc.es/dashboard/enviar-incidencia| requests and problem reporting form]]. Users without access permission will receive an incorrect password error message.
  
 ===== User Manual ===== ===== User Manual =====
Line 98: Line 77:
 Use SSH. Hostnames and ip addresses are: Use SSH. Hostnames and ip addresses are:
  
-  * ctgpgpu2.inv.usc.es - 172.16.242.92:22 +
-  * ctgpgpu3.inv.usc.es - 172.16.242.93:22+
   * ctgpgpu4.inv.usc.es - 172.16.242.201:22   * ctgpgpu4.inv.usc.es - 172.16.242.201:22
   * ctgpgpu5.inv.usc.es - 172.16.242.202:22   * ctgpgpu5.inv.usc.es - 172.16.242.202:22
   * ctgpgpu6.inv.usc.es - 172.16.242.205:22   * ctgpgpu6.inv.usc.es - 172.16.242.205:22
-  * ctgpgpu7.inv.usc.es - 172.16.242.207:22 +  * ctgpgpu9.inv.usc.es - 172.16.242.94:22 
-  * ctgpgpu8.inv.usc.es - 172.16.242.208:22 +  * ctgpgpu10.inv.usc.es - 172.16.242.95:22 
 +  * ctgpgpu11.inv.usc.es - 172.16.242.96:22 
 +  * ctgpgpu12.inv.usc.es - 172.16.242.97:22
 Connection in only possible from inside the CITIUS network. To connect from other places or from the RAI network it is necessary to use the [[https://wiki.citius.usc.es/en:centro:servizos:vpn:start | VPN]] or the [[https://wiki.citius.usc.es/en:centro:servizos:pasarela_ssh|SSH gateway]]. Connection in only possible from inside the CITIUS network. To connect from other places or from the RAI network it is necessary to use the [[https://wiki.citius.usc.es/en:centro:servizos:vpn:start | VPN]] or the [[https://wiki.citius.usc.es/en:centro:servizos:pasarela_ssh|SSH gateway]].