Once you have completed your cloud deployment, these are some of the common post-installation tasks you may need to perform to verify your cloud installation.
The sections below will show you the steps to verify that your Block Storage backend was setup properly.
Perform the following steps to create a volume using Horizon dashboard.
Log into the Horizon dashboard. For more information, see Book “User Guide”, Chapter 3 “Cloud Admin Actions with the Dashboard”.
Choose
› › .On the
tabs, click the button to create a volume.In the
options, enter the required details into the fields and then click the button:Volume Name - This is the name you specify for your volume.
Description (optional) - This is an optional description for the volume.
Type - Select the volume type you have created for your volumes from the drop down.
Size (GB) - Enter the size, in GB, you would like the volume to be.
Availability Zone - You can either leave this at the default option of
or select a specific zone from the drop-down box.The dashboard will then show the volume you have just created.
Perform the following steps to attach a volume to an instance:
Log into the Horizon dashboard. For more information, see Book “User Guide”, Chapter 3 “Cloud Admin Actions with the Dashboard”.
Choose
› › .In the
column, choose the in the drop-down box next to the instance you want to attach the volume to.In the
drop-down, select the volume that you want to attach.Edit the
if necessary.Click
to complete the action.On the
screen, verify that the volume you attached is displayed in the columns.Perform the following steps to detach the volume from instance:
Log into the Horizon dashboard. For more information, see Book “User Guide”, Chapter 3 “Cloud Admin Actions with the Dashboard”.
Choose
› › .Click the check box next to the name of the volume you want to detach.
In the
column, choose the in the drop-down box next to the instance you want to attach the volume to.Click
. A confirmation dialog box appears.Click
to confirm the detachment of the volume from the associated instance.Perform the following steps to delete a volume using Horizon dashboard:
Log into the Horizon dashboard. For more information, see Book “User Guide”, Chapter 3 “Cloud Admin Actions with the Dashboard”.
Choose
› › .In the
column, click next to the volume you would like to delete.To confirm and delete the volume, click
again.Verify that the volume was removed from the
screen.The following procedure shows how to validate that all servers have been added to the Swift rings:
Run the swift-compare-model-rings.yml playbook as follows:
cd ~/scratch/ansible/next/ardana/ansible ansible-playbook -i hosts/verb_hosts swift-compare-model-rings.yml
Search for output similar to the following. Specifically, look at the number of drives that are proposed to be added.
TASK: [swiftlm-ring-supervisor | validate-input-model | Print report] ********* ok: [ardana-cp1-c1-m1-mgmt] => { "var": { "report.stdout_lines": [ "Rings:", " ACCOUNT:", " ring exists", " no device changes", " ring will be rebalanced", " CONTAINER:", " ring exists", " no device changes", " ring will be rebalanced", " OBJECT-0:", " ring exists", " no device changes", " ring will be rebalanced" ] } }
If the text contains "no device changes" then the deploy was successful and no further action is needed.
If more drives need to be added, it indicates that the deploy failed on some nodes and that you restarted the deploy to include those nodes. However, the nodes are not in the Swift rings because enough time has not elapsed to allow the rings to be rebuilt. You have two options to continue:
Repeat the deploy. There are two steps:
Delete the ring builder files as described in Book “Operations Guide”, Chapter 15 “Troubleshooting Issues”, Section 15.6 “Storage Troubleshooting”, Section 15.6.2 “Swift Storage Troubleshooting”, Section 15.6.2.8 “Restarting the Object Storage Deployment”.
Repeat the installation process starting by running the
site.yml
playbook as described in
Section 12.7, “Deploying the Cloud”.
Rebalance the rings several times until all drives are incorporated in the rings. This process may take several hours to complete (because you need to wait one hour between each rebalance). The steps are as follows:
Change the min-part-hours to 1 hour. See Book “Operations Guide”, Chapter 8 “Managing Object Storage”, Section 8.5 “Managing Swift Rings”, Section 8.5.7 “Changing min-part-hours in Swift”.
Use the "First phase of ring rebalance" and "Final rebalance phase" as described in Book “Operations Guide”, Chapter 8 “Managing Object Storage”, Section 8.5 “Managing Swift Rings”, Section 8.5.5 “Applying Input Model Changes to Existing Rings”. The “Weight change phase of ring rebalance” does not apply because you have not set the weight-step attribute at this stage.
Set the min-part-hours to the recommended 16 hours as described in Book “Operations Guide”, Chapter 8 “Managing Object Storage”, Section 8.5 “Managing Swift Rings”, Section 8.5.7 “Changing min-part-hours in Swift”.
If you receive errors during the validation, see Book “Operations Guide”, Chapter 15 “Troubleshooting Issues”, Section 15.6 “Storage Troubleshooting”, Section 15.6.2 “Swift Storage Troubleshooting”, Section 15.6.2.3 “Interpreting Swift Input Model Validation Errors”.
For information about verifying the operations, see Book “Operations Guide”, Chapter 8 “Managing Object Storage”, Section 8.1 “Running the Swift Dispersion Report”.
To create a Compute instance, you need to obtain an image that you can use. The Cloud Lifecycle Manager provides an Ansible playbook that will download a CirrOS Linux image, and then upload it as a public image to your image repository for use across your projects.
Use the following command to run this playbook:
cd ~/scratch/ansible/next/ardana/ansible ansible-playbook -i hosts/verb_hosts glance-cloud-configure.yml -e proxy=<PROXY>
The table below shows the optional switch that you can use as part of this playbook to specify environment-specific information:
Switch | Description |
---|---|
|
Optional. If your environment requires a proxy for the internet, use this switch to specify the proxy information. |
OpenStack has created a guide to show you how to obtain, create, and modify images that will be compatible with your cloud:
You can use the GlanceClient on a machine accessible to your cloud or on your Cloud Lifecycle Manager where it is automatically installed.
The GlanceClient allows you to create, update, list, and delete images as
well as manage your image member lists, which allows you to share access to
images across multiple tenants. As with most of the OpenStack CLI tools, you
can use the glance help
command to get a full list of
commands as well as their syntax.
If you would like to use the --copy-from
option when
creating an image, you will need to have your Administrator enable the http
store in your environment using the instructions outlined at
Book “Operations Guide”, Chapter 5 “Managing Compute”, Section 5.6 “Configuring the Image Service”, Section 5.6.2 “Allowing the Glance copy-from option in your environment”.
You must have an external network set up to allow your Compute instances to reach the internet. There are multiple methods you can use to create this external network and we provide two of them here. The HPE Helion OpenStack installer provides an Ansible playbook that will create this network for use across your projects. We also show you how to create this network via the command line tool from your Cloud Lifecycle Manager.
This playbook will query the Networking service for an existing external
network, and then create a new one if you do not already have one. The
resulting external network will have the name ext-net
with a subnet matching the CIDR you specify in the command below.
If you need to specify more granularity, for example specifying an allocation pool for the subnet then you should utilize the Section 28.4.2, “Using the NeutronClient CLI”.
cd ~/scratch/ansible/next/ardana/ansible ansible-playbook -i hosts/verb_hosts neutron-cloud-configure.yml -e EXT_NET_CIDR=<CIDR>
The table below shows the optional switch that you can use as part of this playbook to specify environment-specific information:
Switch | Description |
---|---|
|
Optional. You can use this switch to specify the external network CIDR. If you choose not to use this switch, or use a wrong value, the VMs will not be accessible over the network.
This CIDR will be from the NoteIf this option is not defined the default value is "172.31.0.0/16" |
For more granularity you can utilize the Neutron command line tool to create your external network.
Log in to the Cloud Lifecycle Manager.
Source the Admin credentials:
source ~/service.osrc
Create the external network and then the subnet using these commands below.
Creating the network:
neutron net-create --router:external <external-network-name>
Creating the subnet:
neutron subnet-create <external-network-name> <CIDR> --gateway <gateway> \ --allocation-pool start=<IP_start>,end=<IP_end> [--disable-dhcp]
Where:
Value | Description |
---|---|
external-network-name |
This is the name given to your external network. This is a unique
value that you will choose. The value |
CIDR |
You can use this switch to specify the external network CIDR. If you choose not to use this switch, or use a wrong value, the VMs will not be accessible over the network. This CIDR will be from the EXTERNAL VM network. |
--gateway |
Optional switch to specify the gateway IP for your subnet. If this is not included then it will choose the first available IP. |
--allocation-pool start end |
Optional switch to specify a start and end IP address to use as the allocation pool for this subnet. |
--disable-dhcp |
Optional switch if you want to disable DHCP on this subnet. If this is not specified then DHCP will be enabled. |
Once the external network is created, users can create a Private Network to complete their networking setup. For instructions, see Book “User Guide”, Chapter 8 “Creating a Private Network”.