You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
we would like to request a small GPU python kernel for UC3 / UC4 under the EOX Lab environment which will further test headless execution functionality and the the provisioning of GPU resources to the UCs. one multi-GPU machine with e.g. 8x A100 would needed for testing and semi-production.
The text was updated successfully, but these errors were encountered:
"The node_purpose can either have the value user to use a CPU node with 2 CPUs and 8 GB memory or userg1 to use a GPU node with 4 CPUs, 16 GB memory, and a Tesla T4 GPU with 16 GB memory."
The node_purpose takes only two values user or userg1 (if we need GPUs). We are wondering if there is another value/option with higher GPU configuration?
Correct, node_purpose currently supports two values: user for a CPU instance and userg1 for a GPU one. For the GPU one the AWS EC2 type g4dn.xlarge is used providing one NVIDIA T4 GPU. This type costs less than 1€ per hour.
I believe we can configure an additional instance type like for example a p4d.24xlarge providing 8 NVIDIA Tesla A100 GPUs but the costs are more than 40€ per hour. Please also note that we have no particular experience with multi-GPU machines.
Anyway, I'll request the configuration from our DevOps team and keep you informed.
All the headless endpoints can be started either with:
node_purpose "user" (for regular CPU)
node_purpose "userg1" (for single GPU) or
if node_purpose is not passed then "userg1" is default
In addition the smallest Multi GPU VM available on eu-central-1 g4dn.12xlarge (4 x NVIDIA T4 16 GiB) -> $4.89 per hour on "userg2" has been configured. Only UC3 (eurodatacube17) and UC4 (eurodatacube18) are whitelisted for using "userg2".
we would like to request a small GPU python kernel for UC3 / UC4 under the EOX Lab environment which will further test headless execution functionality and the the provisioning of GPU resources to the UCs. one multi-GPU machine with e.g. 8x A100 would needed for testing and semi-production.
The text was updated successfully, but these errors were encountered: