6. Managing GPUs¶
CloudVeneto provides Unipd Physics Dept. and INFN Padova users with some GPUs (Graphics Processing Units). These are:
- 4 GPU Nvidia V100
- 4 GPU Nvidia Tesla T4
- 1 GPU Nvidia Quadro RTX 6000
- 2 GPU Nvidia TITAN Xp
- 1 GPU Nvidia GeForce GTX TITAN
Using a CloudVeneto GPU means accessing a virtual machine which has full access and direct control of such GPU device.
GPU instances, i.e. virtual machines which have access to one or more GPUs can be created only from the HPC-Physics project. The only exception is is for the T4 GPUs that are usable also from the PhysicsOfData-students project.
So, first of all, you need to request the affiliation to such project (see Apply for other projects for the relevant instructions).
6.1. Reserving a GPU¶
Before using a GPU, you need to reserve it. This can be done using a reservation system integrated in the dashboard.
Using the Dashboard, click on GPU Booking Calendar.
Let’s suppose that you want to reserve a Tesla V100 GPU from Oct 5 to Oct 9.
Move the desired GPU to the first day of the reservation (October 5, in our example)
Using the mouse, you can then “enlarge” your reservation till the desired last day (October 9, in our example)
You may also associate a comment for this reservation (by clicking on it). The message can be seen by the other users.
Please note that a reservation can be at most 15 days long.
To delete a reservation, you simply need to move it to the trash bin.
The reservation system that has been just described, is visible only to the projects that have access to the GPUs (i.e. the HPC-Physics project and, just for the T4 GPUs, the PhysicsOfData-students project)
6.2. Creating a GPU instance¶
The instructions to create a GPU instance are the very same for the creation of a ‘standard’ virtual machine (see Creating Virtual Machines). You will only have to pay attention to use one of these special flavors:
Flavor for an instance with 1 GPU Nvidia V100, 18 VCPUs, 56 GB of RAM, 20 GB of ephemeral root disk space, 75 GB of extra ephemeral disk space.
Flavor for an instance with 2 GPU Nvidia V100, 36 VCPUs, 112 GB of RAM, 20 GB of ephemeral root disk space, 170 GB of extra ephemeral disk space.
Flavor for an instance with 4 GPU Nvidia V100, 72 VCPUs, 224 GB of RAM, 20 GB of ephemeral root disk space, 360 GB of extra ephemeral disk space.
Flavor for an instance with 1 GPU Nvidia T4, 15 VCPUs, 90 GB of RAM, 20 GB of ephemeral root disk space, 500 GB of extra ephemeral disk space.
Flavor for an instance with 2 GPUs Nvidia T4, 30 VCPUs, 180 GB of RAM, 20 GB of ephemeral root disk space, 1400 GB of extra ephemeral disk space.
Flavor for an instance with 4 GPUs Nvidia T4, 60 VCPUs, 360 GB of RAM, 20 GB of ephemeral root disk space, 2800 GB of extra ephemeral disk space.
Flavor for an instance with 1 GPU Nvidia Quadro RTX 6000, 8 VCPUs, 40 GB of RAM, 20 GB of ephemeral root disk space, 500 GB of extra ephemeral disk space.
Flavor for an instance with 1 GPU Nvidia Titan Xp, 8 VCPUs, 40 GB of RAM, 20 GB of ephemeral root disk space, 400 GB of extra ephemeral disk space.
Flavor for an instance with 2 GPUs Nvidia Titan Xp, 16 VCPUs, 80 GB of RAM, 20 GB of ephemeral root disk space, 800 GB of extra ephemeral disk space.
Flavor for an instance with 1 GPU Nvidia GeForce GTX TITAN, 4 VCPUs, 20 GB of RAM, 20 GB of ephemeral root disk space, 200 GB of extra ephemeral disk space.
When you snapshot an instance created using one of such flavors, please consider that only the root disk is saved. The content of the extra ephemeral disk is not saved !
For historical reasons there are still some instances created using flavors with a large root disk: since the size limit for images and snapshots is 25 GB, such instances cannot be snapshotted.
Before allocating one or more GPUs, please register such allocation using the reservation system described in the previous section.
6.2.1. Images for GPU instances¶
These instructions explain how to install CUDA toolkit and the relevant drivers.
For better performance, we suggest to create images:
in raw format
setting the properties hw_disk_bus=scsi and hw_scsi_model=virtio-scsi, i.e., using the command line tool:# glance image-update --property hw_disk_bus=scsi <image-id> # glance image-update --property hw_scsi_model=virtio-scsi <image-id>
Unfortunately it is not straightforward to see which GPUs are being used and which ones are available using the CloudVeneto Openstack dashboard.
You can refer to this page for such information (please note that this page is updated every 30 minutes).
Please consider the following policies when using GPU instances:
- Since there is a high request to use GPUs, please delete your instance as soon as you don’t need it anymore. This is because virtual machines, even if idle or in shutdown state, allocate resources (GPUs in particular) which therefore aren’t available to other users.
- Once activated, your virtual instance is managed by you.
- Before allocation one or more GPUs, please register such allocation as explained here. Instances for which there isn’t a reservation can be deleted by the Cloud administrators.
- Please don’t reserve the GPU(s) for long (i.e > 1 week) periods.