Alert.png The wiki is deprecated and due to be decommissioned by the end of September 2022.
The content is being migrated to other supports, new updates will be ignored and lost.
If needed you can get in touch with EGI SDIS team using operations @ egi.eu.

Difference between revisions of "GPGPU-OpenNebula"

From EGIWiki
Jump to navigation Jump to search
(Created page with "{{Template:EGI-Engage menubar}} {{TOC_right}} = Objective = To provide testing Cloud site based on OpenNebula middleware for testing GPGPU support. = Current status = IISAS-...")
 
Line 52: Line 52:
  EGI AppDB: https://appdb.egi.eu/store/site/iisas-nebula
  EGI AppDB: https://appdb.egi.eu/store/site/iisas-nebula
  Supported VOs: acc-comp.egi.eu, ops, dteam
  Supported VOs: acc-comp.egi.eu, ops, dteam
= How to use IISAS-Nebula site =
;Join [[Accelerated_computing_VO]]
:VO acc-comp.egi.eu is dedicated for users to develop and test applications/VMs that use GPGPU or other types of accelerated computing.
;Install rOCCI client
:More information about installation and using of rOCCI CLI can be found at [[HOWTO11_How_to_use_the_rOCCI_Client]]
;Get RFC proxy certificate from acc-comp.egi.eu VOMS server
$ voms-proxy-init --voms acc-comp.egi.eu -rfc
;Choose a suitable flavor from the list above
:Alternatively you can list the available resource flavors using OCCI client:
$ occi --endpoint https://nebula2.ui.savba.sk:11443/ --auth x509 --user-cred $X509_USER_PROXY --voms \
        --action describe --resource resource_tpl
;Choose a suitable image from the list of supported Virtual Appliance images
:The up-to-date list can be found at [https://appdb.egi.eu/store/site/iisas-nebula EGI AppDB] or using OCCI client:
$ occi --endpoint https://nebula2.ui.savba.sk:11443/ --auth x509 --user-cred $X509_USER_PROXY --voms \
        --action describe --resource os_tpl
;Create a VM with the selected image, flavor and SSH keypair using OCCI command:
$ occi --endpoint https://nebula2.ui.savba.sk:11443/ --auth x509 --user-cred $X509_USER_PROXY --voms \
        --action create --resource compute \
        --mixin os_tpl#uuid_egi_centos_7_8 \
        --mixin resource_tpl#small_gpu \
        --attribute occi.core.title="Testing GPU" \
        --context user_data="file://$PWD/tmpfedcloud.login"
:The command should print the URL ID of your new VM
;Find out IP address assigned to your new VM
$ occi --endpoint https://nebula2.ui.savba.sk:11443/ --auth x509 --user-cred $X509_USER_PROXY --voms \
        --action describe --resource $VM_ID_URL | grep occi.networkinterface.address
;Log in the VM with your private key
$ ssh -i tmpfedcloud testadm@$VM_PUBLIC_IP
;Install Nvidia drivers and CUDA toolkit
:Example installation of CUDA 8.0 for CentOS 7:
[testadm@localhost ~]$ sudo yum -y update ; sudo reboot
[testadm@localhost ~]$ sudo yum -y install kernel-devel-$(uname -r) kernel-headers-$(uname -r)
[testadm@localhost ~]$ sudo yum -y install http://developer.download.nvidia.com/compute/cuda/repos/rhel7/x86_64/cuda-repo-rhel7-8.0.61-1.x86_64.rpm
[testadm@localhost ~]$ sudo yum install cuda
;Delete the VM after you finish working with it
$ occi --endpoint https://nebula2.ui.savba.sk:11443/ --auth x509 --user-cred $X509_USER_PROXY --voms \
        --action delete --resource $VM_ID_URL

Revision as of 14:44, 12 April 2017

EGI-Engage project: Main page WP1(NA1) WP3(JRA1) WP5(SA1) PMB Deliverables and Milestones Quality Plan Risk Plan Data Plan
Roles and
responsibilities
WP2(NA2) WP4(JRA2) WP6(SA2) AMB Software and services Metrics Project Office Procedures



Objective

To provide testing Cloud site based on OpenNebula middleware for testing GPGPU support.

Current status

IISAS-Nebula site has been integrated to EGI Federated Cloud and is accessible using acc-comp.egi.eu VO.

HW configuration:

Management services: OpenNebula Cloud controller and Site BDII in virtual servers
IBM System x3250 M5, 1x Intel(R) Xeon(R) CPU E3-1241 v3 @ 3.50GHz, 16 RAM, 1TB Disk

1 computing node: IBM dx360 M4 server with two NVIDIA Tesla K20 accelerators.
CentOS 7 with KVM/QEMU, PCI passthrough virtualization of GPU cards.

2.8TB block storage via NFS

SW configuration:

Base OS: CentOS 7
Hypervisor: KVM
Middleware: OpenNebula 5.0.2
OCCI server: rOCCI-server 2.0.0

GPU-enabled flavors:

extra_large_2gpu      Extra Large Instance - 8 cores and 8 GB RAM + 2 GPU Nvidia K20m
extra_large_gpu       Extra Large Instance - 8 cores and 8 GB RAM + 1 GPU Nvidia K20m
goliath_2gpu          Goliath Instance - 14 cores and 56 GB RAM + 2 GPU Nvidia K20m
goliath_gpu           Goliath Instance - 14 cores and 56 GB RAM + 1 GPU Nvidia K20m
large_2gpu            Large Instance - 4 cores and 4 GB RAM + 2 GPU Nvidia K20m
large_gpu             Large Instance - 4 cores and 4 GB RAM + 1 GPU Nvidia K20m
mammoth_2gpu          Mammoth Instance - 14 cores and 32 GB RAM + 2 GPU Nvidia K20m
mammoth_gpu           Mammoth Instance - 14 cores and 32 GB RAM + 1 GPU Nvidia K20m
medium_2gpu           Medium Instance - 2 cores and 2 GB RAM + 2 GPU Nvidia K20m
medium_gpu            Medium Instance - 2 cores and 2 GB RAM + 1 GPU Nvidia K20m
mem_extra_large_2gpu  Extra Large Instance - 8 cores and 32 GB RAM + 2 GPU Nvidia K20m
mem_extra_large_gpu   Extra Large Instance - 8 cores and 32 GB RAM + 1 GPU Nvidia K20m
mem_large_2gpu        Large Instance - 4 cores and 16 GB RAM + 2 GPU Nvidia K20m
mem_large_gpu         Large Instance - 4 cores and 16 GB RAM + 1 GPU Nvidia K20m
mem_medium_2gpu       Medium Instance - 2 cores and 8 GB RAM + 2 GPU Nvidia K20m
mem_medium_gpu        Medium Instance - 2 cores and 8 GB RAM + 1 GPU Nvidia K20m
mem_small_2gpu        Small Instance - 1 core and 4 GB RAM + 2 GPU Nvidia K20m
mem_small_gpu         Small Instance - 1 core and 4 GB RAM + 1 GPU Nvidia K20m
small_2gpu            Small Instance - 1 core and 1 GB RAM + 2 GPU Nvidia K20m
small_gpu             Small Instance - 1 core and 1 GB RAM + 1 GPU Nvidia K20m

EGI federated cloud configuration:

GOCDB: IISAS-Nebula, https://goc.egi.eu/portal/index.php?Page_Type=Site&id=1785
ARGO monitoring: http://argo.egi.eu/lavoisier/status_report-sf?site=IISAS-Nebula&report=Critical&accept=html
OCCI endpoint: https://nebula2.ui.savba.sk:11443/
EGI AppDB: https://appdb.egi.eu/store/site/iisas-nebula
Supported VOs: acc-comp.egi.eu, ops, dteam

How to use IISAS-Nebula site

Join Accelerated_computing_VO
VO acc-comp.egi.eu is dedicated for users to develop and test applications/VMs that use GPGPU or other types of accelerated computing.
Install rOCCI client
More information about installation and using of rOCCI CLI can be found at HOWTO11_How_to_use_the_rOCCI_Client
Get RFC proxy certificate from acc-comp.egi.eu VOMS server
$ voms-proxy-init --voms acc-comp.egi.eu -rfc
Choose a suitable flavor from the list above
Alternatively you can list the available resource flavors using OCCI client:
$ occi --endpoint https://nebula2.ui.savba.sk:11443/ --auth x509 --user-cred $X509_USER_PROXY --voms \
       --action describe --resource resource_tpl
Choose a suitable image from the list of supported Virtual Appliance images
The up-to-date list can be found at EGI AppDB or using OCCI client:
$ occi --endpoint https://nebula2.ui.savba.sk:11443/ --auth x509 --user-cred $X509_USER_PROXY --voms \
       --action describe --resource os_tpl
Create a VM with the selected image, flavor and SSH keypair using OCCI command
$ occi --endpoint https://nebula2.ui.savba.sk:11443/ --auth x509 --user-cred $X509_USER_PROXY --voms \
       --action create --resource compute \
       --mixin os_tpl#uuid_egi_centos_7_8 \
       --mixin resource_tpl#small_gpu \
       --attribute occi.core.title="Testing GPU" \
       --context user_data="file://$PWD/tmpfedcloud.login"
The command should print the URL ID of your new VM
Find out IP address assigned to your new VM
$ occi --endpoint https://nebula2.ui.savba.sk:11443/ --auth x509 --user-cred $X509_USER_PROXY --voms \
       --action describe --resource $VM_ID_URL | grep occi.networkinterface.address
Log in the VM with your private key
$ ssh -i tmpfedcloud testadm@$VM_PUBLIC_IP
Install Nvidia drivers and CUDA toolkit
Example installation of CUDA 8.0 for CentOS 7:
[testadm@localhost ~]$ sudo yum -y update ; sudo reboot
[testadm@localhost ~]$ sudo yum -y install kernel-devel-$(uname -r) kernel-headers-$(uname -r)
[testadm@localhost ~]$ sudo yum -y install http://developer.download.nvidia.com/compute/cuda/repos/rhel7/x86_64/cuda-repo-rhel7-8.0.61-1.x86_64.rpm
[testadm@localhost ~]$ sudo yum install cuda
Delete the VM after you finish working with it
$ occi --endpoint https://nebula2.ui.savba.sk:11443/ --auth x509 --user-cred $X509_USER_PROXY --voms \
       --action delete --resource $VM_ID_URL