diff --git a/doc/training-guides/basic-install-guide/lab_control-node.xml b/doc/training-guides/basic-install-guide/lab_control-node.xml
index 0c6fe8e6..36f90c2e 100644
--- a/doc/training-guides/basic-install-guide/lab_control-node.xml
+++ b/doc/training-guides/basic-install-guide/lab_control-node.xml
@@ -68,20 +68,20 @@ iface lo inet loopback
# The primary network interface - Virtual Box NAT connection
# (Virtual Box Network Adapter 3)
-auto eth2
-iface eth2 inet dhcp
+auto eth0
+iface eth0 inet dhcp
# Virtual Box vboxnet0 - OpenStack Management Network
# (Virtual Box Network Adapter 1)
-auto eth0
-iface eth0 inet static
+auto eth1
+iface eth1 inet static
address 10.10.10.51
netmask 255.255.255.0
# Virtual Box vboxnet2 - for exposing OpenStack API over external network
# (Virtual Box Network Adapter 2)
-auto eth1
-iface eth1 inet static
+auto eth2
+iface eth2 inet static
address 192.168.100.51
netmask 255.255.255.0
@@ -188,29 +188,6 @@ netmask 255.255.255.0
Other
-
-
- Install other services:
- # apt-get install -y vlan bridge-utils
-
-
- Enable IP_Forwarding:
- # sed -i 's/#net.ipv4.ip_forward=1/net.ipv4.ip_forward=1/' /etc/sysctl.conf
-
-
- Also add the following two lines
- into/etc/sysctl.conf:
- net.ipv4.conf.all.rp_filter=0
- net.ipv4.conf.default.rp_filter=0
-
-
- To save you from reboot, perform the following
- # sysctl net.ipv4.ip_forward=1
- # sysctl net.ipv4.conf.all.rp_filter=0
- # sysctl net.ipv4.conf.default.rp_filter=0
- # sysctl -p
-
-
Keystone
Keystone is an OpenStack project that provides Identity,
Token, Catalog and Policy services for use specifically by
@@ -336,7 +313,13 @@ admin_password = service_pass
Update the
/etc/glance/glance-api.conf:
- sql_connection = mysql://glanceUser:glancePass@10.10.10.51/glance
+
+[DEFAULT]
+rpc_backend = rabbit
+rabbit_host = 10.10.10.51
+
+[database]
+sql_connection = mysql://glanceUser:glancePass@10.10.10.51/glance
[keystone_authtoken]
auth_host = 10.10.10.51
auth_port = 35357
@@ -351,7 +334,8 @@ flavor = keystone
Update the
/etc/glance/glance-registry.conf:
- sql_connection = mysql://glanceUser:glancePass@10.10.10.51/glance
+ [database]
+sql_connection = mysql://glanceUser:glancePass@10.10.10.51/glance
[keystone_authtoken]
auth_host = 10.10.10.51
auth_port = 35357
@@ -382,7 +366,74 @@ flavor = keystone
$ glance image-list
- Neutron
+ Nova
+ Nova is the project name for OpenStack Compute, a cloud
+ computing fabric controller, the main part of an IaaS system.
+ Individuals and organizations can use Nova to host and manage
+ their own cloud computing systems. Nova originated as a project
+ out of NASA Ames Research Laboratory.
+ Nova is written with the following design guidelines in
+ mind:
+ Install nova components:
+
+ # apt-get install -y nova-api nova-cert nova-conductor nova-consoleauth nova-novncproxy nova-scheduler python-novaclient
+
+
+
+ Edit /etc/nova/nova.conf
+ [database]
+connection = mysql://novaUser:novaPass@10.10.10.51/nova
+
+[DEFAULT]
+rpc_backend = rabbit
+rabbit_host = 10.10.10.51
+my_ip = 10.10.10.51
+vncserver_listen = 10.10.10.51
+vncserver_proxyclient_address = 10.10.10.51
+auth_strategy = keystone
+
+network_api_class = nova.network.neutronv2.api.API
+neutron_url = http://10.10.10.51:9696
+neutron_auth_strategy = keystone
+neutron_admin_tenant_name = service
+neutron_admin_username = neutron
+neutron_admin_password = service_pass
+neutron_admin_auth_url = http://10.10.10.51:35357/v2.0
+linuxnet_interface_driver = nova.network.linux_net.LinuxOVSInterfaceDriver
+firewall_driver = nova.virt.firewall.NoopFirewallDriver
+security_group_api = neutron
+
+[keystone_authtoken]
+auth_uri = http://10.10.10.51:5000
+auth_host = controller
+auth_port = 35357
+auth_protocol = http
+admin_tenant_name = service
+admin_user = nova
+admin_password = service_pass
+
+
+
+ Synchronize your database:
+ # nova-manage db sync
+
+
+ Restart nova-* services (all nova services):
+ #service nova-api restart
+ #service nova-cert restart
+ #service nova-consoleauth restart
+ #service nova-scheduler restart
+ #service nova-conductor restart
+ #service nova-novncproxy restart
+
+
+ Check for the smiling faces on
+ nova-* services to confirm your
+ installation:
+ # nova-manage service list
+
+
+Neutron
Neutron is an OpenStack project to provide “network
connectivity as a service" between interface devices (e.g., vNICs)
managed by other OpenStack services (e.g., nova).
@@ -390,26 +441,23 @@ flavor = keystone
Install the Neutron Server and the Open vSwitch package
collection:
- # apt-get install -y neutron-server
+ # apt-get install -y neutron-server neutron-plugin-ml2
Edit the
- /etc/neutron/plugins/openvswitch/ovs_neutron_plugin.ini:
- [database]
-connection = mysql://neutronUser:neutronPass@10.10.10.51/neutron
+ /etc/neutron/plugins/ml2/ml2_conf.ini:
+ [ml2]
+type_drivers = gre
+tenant_network_types = gre
+mechanism_drivers = openvswitch
-#Under the OVS section
-[ovs]
-tenant_network_type = gre
+[ml2_type_gre]
tunnel_id_ranges = 1:1000
-enable_tunneling = True
-[agent]
-tunnel_types = gre
-#Firewall driver for realizing neutron security group function
[securitygroup]
firewall_driver =
-neutron.agent.linux.iptables_firewall.OVSHybridIptablesFirewallDriver
+neutron.agent.linux.iptables_firewall.OVSHybridIptablesFirewallDriver
+enable_security_group = True
Edit the
@@ -428,7 +476,18 @@ admin_password = service_pass
Edit the
/etc/neutron/neutron.conf:
- rabbit_host = 10.10.10.51
+ [DEFAULT]
+auth_strategy = keystone
+rpc_backend = neutron.openstack.common.rpc.impl_kombu
+rabbit_host = controller
+notify_nova_on_port_status_changes = True
+notify_nova_on_port_data_changes = True
+nova_url = http://controller:8774/v2
+nova_admin_username = nova
+nova_admin_tenant_id = SERVICE_TENANT_ID
+nova_admin_password = NOVA_PASS
+nova_admin_auth_url = http://controller:35357/v2.0
+
[keystone_authtoken]
auth_host = 10.10.10.51
auth_port = 35357
@@ -442,210 +501,53 @@ signing_dir = /var/lib/neutron/keystone-signing
connection = mysql://neutronUser:neutronPass@10.10.10.51/neutron
+ Restart Nova Services
+ # service nova-api restart
+ # service nova-scheduler restart
+ # service nova-conductor restart
Restart Neutron services:
# service neutron-server restart
- Nova
- Nova is the project name for OpenStack Compute, a cloud
- computing fabric controller, the main part of an IaaS system.
- Individuals and organizations can use Nova to host and manage
- their own cloud computing systems. Nova originated as a project
- out of NASA Ames Research Laboratory.
- Nova is written with the following design guidelines in
- mind:
-
-
- Component based architecture: Quickly adds new
- behaviors.
-
-
- Highly available: Scales to very serious workloads.
-
-
- Fault-Tolerant: Isolated processes avoid cascading
- failures.
-
-
- Recoverable: Failures should be easy to diagnose, debug,
- and rectify.
-
-
- Open standards: Be a reference implementation for a
- community-driven api.
-
-
- API compatibility: Nova strives to be API-compatible with
- popular systems like Amazon EC2.
-
-
- Install nova components:
- # apt-get install -y nova-novncproxy novnc nova-api nova-ajax-console-proxy nova-cert nova-conductor nova-consoleauth nova-doc nova-scheduler python-novaclient
-
-
- Edit /etc/nova/api-paste.ini
- [filter:authtoken]
-paste.filter_factory = keystoneclient.middleware.auth_token:filter_factory
-auth_host = 10.10.10.51
-auth_port = 35357
-auth_protocol = http
-admin_tenant_name = service
-admin_user = nova
-admin_password = service_pass
-signing_dir = /tmp/keystone-signing-nova
-# Workaround for https://bugs.launchpad.net/nova/+bug/1154809
-auth_version = v2.0
-
-
- Edit /etc/nova/nova.conf
- [DEFAULT]
-logdir=/var/log/nova
-state_path=/var/lib/nova
-lock_path=/run/lock/nova
-verbose=True
-api_paste_config=/etc/nova/api-paste.ini
-compute_scheduler_driver=nova.scheduler.simple.SimpleScheduler
-rabbit_host=10.10.10.51
-nova_url=http://10.10.10.51:8774/v1.1/
-sql_connection=mysql://novaUser:novaPass@10.10.10.51/nova
-root_helper=sudo nova-rootwrap /etc/nova/rootwrap.conf
-
-# Auth
-use_deprecated_auth=false
-auth_strategy=keystone
-
-# Imaging service
-glance_api_servers=10.10.10.51:9292
-image_service=nova.image.glance.GlanceImageService
-
-# Vnc configuration
-novnc_enabled=true
-novncproxy_base_url=http://192.168.1.51:6080/vnc_auto.html
-novncproxy_port=6080
-vncserver_proxyclient_address=10.10.10.51
-vncserver_listen=0.0.0.0
-
-# Network settings
-network_api_class=nova.network.neutronv2.api.API
-neutron_url=http://10.10.10.51:9696
-neutron_auth_strategy=keystone
-neutron_admin_tenant_name=service
-neutron_admin_username=neutron
-neutron_admin_password=service_pass
-neutron_admin_auth_url=http://10.10.10.51:35357/v2.0
-libvirt_vif_driver=nova.virt.libvirt.vif.LibvirtHybridOVSBridgeDriver
-linuxnet_interface_driver=nova.network.linux_net.LinuxOVSInterfaceDriver
-
-#If you want Neutron + Nova Security groups
-firewall_driver=nova.virt.firewall.NoopFirewallDriver
-security_group_api=neutron
-#If you want Nova Security groups only, comment the two lines above and
-uncomment line -1-.
-#-1-firewall_driver=nova.virt.libvirt.firewall.IptablesFirewallDriver
-
-#Metadata
-service_neutron_metadata_proxy = True
-neutron_metadata_proxy_shared_secret = helloOpenStack
-
-# Compute #
-compute_driver=libvirt.LibvirtDriver
-
-# Cinder #
-volume_api_class=nova.volume.cinder.API
-osapi_volume_listen_port=5900
-
-
- Synchronize your database:
- # nova-manage db sync
-
-
- Restart nova-* services (all nova services):
- # cd /etc/init.d/; for i in $( ls nova-* ); do service $i restart; done
-
-
- Check for the smiling faces on nova-* services to confirm your
- installation:
- # nova-manage service list
-
-
Cinder
Cinder is an OpenStack project that provides “block storage as a
service”.
-
- Component based architecture: Quickly adds new
- behavior.
-
-
- Highly available: Scales to very serious workloads.
-
-
- Fault-Tolerant: Isolated processes avoid cascading
- failures.
-
-
- Recoverable: Failures should be easy to diagnose, debug
- and rectify.
-
-
- Open standards: Be a reference implementation for a
- community-driven API.
-
-
- API compatibility: Cinder strives to be API-compatible
- with popular systems like Amazon EC2.
-
Install Cinder components:
- # apt-get install -y cinder-api cinder-scheduler cinder-volume iscsitarget open-iscsi iscsitarget-dkms
+ # apt-get install -y cinder-api cinder-scheduler cinder-volume lvm2
- Configure the iSCSI services:
- # sed -i 's/false/true/g' /etc/default/iscsitarget
-
-
- Restart the services:
- # service iscsitarget start
- # service open-iscsi start
-
-
- Edit
- /etc/cinder/api-paste.ini:
- [filter:authtoken]
-paste.filter_factory = keystoneclient.middleware.auth_token:filter_factory
-service_protocol = http
-service_host = 192.168.100.51
-service_port = 5000
+ Edit /etc/cinder/cinder.conf:
+ [database]
+sql_connection = mysql://cinderUser:cinderPass@10.10.10.51/cinder
+
+[keystone_authtoken]
+auth_uri = http://10.10.10.51:5000
auth_host = 10.10.10.51
auth_port = 35357
auth_protocol = http
admin_tenant_name = service
admin_user = cinder
admin_password = service_pass
-signing_dir = /var/lib/cinder
-
-
- Edit /etc/cinder/cinder.conf:
- [DEFAULT]
-rootwrap_config=/etc/cinder/rootwrap.conf
-sql_connection = mysql://cinderUser:cinderPass@10.10.10.51/cinder
-api_paste_config = /etc/cinder/api-paste.ini
-iscsi_helper=ietadm
-volume_name_template = volume-%s
-volume_group = cinder-volumes
-verbose = True
-auth_strategy = keystone
-iscsi_ip_address=10.10.10.51
+
+[DEFAULT]
rpc_backend = cinder.openstack.common.rpc.impl_kombu
rabbit_host = 10.10.10.51
-rabbit_port = 5672
+rabbit_port = 5672
+rabbit_userid = guest
+
Then, synchronize Cinder database:
# cinder-manage db sync
+
+ Restart Cinder Services
+ # service cinder-scheduler restart
+ # service cinder-api restart
+
Finally, create a volume group and name it
cinder-volumes:
@@ -673,14 +575,6 @@ rabbit_port = 5672
down.
-
- Restart the Cinder services:
- # cd /etc/init.d/; for i in $( ls cinder-* ); do service $i restart; done
-
-
- Verify that the Cinder services are running:
- # cd /etc/init.d/; for i in $( ls cinder-* ); do service $i status; done
-
Horizon
Horizon is the canonical implementation of OpenStack’s