Merge "Update armada multinode manifests"
This commit is contained in:
commit
550293a400
@ -20,7 +20,19 @@ source ./tools/deployment/armada/generate-passwords.sh
|
|||||||
: ${OSH_INFRA_PATH:="../openstack-helm-infra"}
|
: ${OSH_INFRA_PATH:="../openstack-helm-infra"}
|
||||||
: ${OSH_PATH:="./"}
|
: ${OSH_PATH:="./"}
|
||||||
|
|
||||||
|
[ -s /tmp/ceph-fs-uuid.txt ] || uuidgen > /tmp/ceph-fs-uuid.txt
|
||||||
|
#NOTE(portdirect): to use RBD devices with Ubuntu kernels < 4.5 this
|
||||||
|
# should be set to 'hammer'
|
||||||
|
. /etc/os-release
|
||||||
|
if [ "x${ID}" == "xubuntu" ] && \
|
||||||
|
[ "$(uname -r | awk -F "." '{ print $2 }')" -lt "5" ]; then
|
||||||
|
export CRUSH_TUNABLES=hammer
|
||||||
|
else
|
||||||
|
export CRUSH_TUNABLES=null
|
||||||
|
fi
|
||||||
|
|
||||||
export CEPH_NETWORK=$(./tools/deployment/multinode/kube-node-subnet.sh)
|
export CEPH_NETWORK=$(./tools/deployment/multinode/kube-node-subnet.sh)
|
||||||
|
export CEPH_FS_ID="$(cat /tmp/ceph-fs-uuid.txt)"
|
||||||
export TUNNEL_DEVICE=$(ip -4 route list 0/0 | awk '{ print $5; exit }')
|
export TUNNEL_DEVICE=$(ip -4 route list 0/0 | awk '{ print $5; exit }')
|
||||||
export OSH_INFRA_PATH
|
export OSH_INFRA_PATH
|
||||||
export OSH_PATH
|
export OSH_PATH
|
||||||
|
@ -73,32 +73,27 @@ data:
|
|||||||
namespace: ceph
|
namespace: ceph
|
||||||
ceph_mon:
|
ceph_mon:
|
||||||
namespace: ceph
|
namespace: ceph
|
||||||
fluentd:
|
|
||||||
namespace: osh-infra
|
|
||||||
network:
|
network:
|
||||||
public: ${CEPH_NETWORK}
|
public: ${CEPH_NETWORK}
|
||||||
cluster: ${CEPH_NETWORK}
|
cluster: ${CEPH_NETWORK}
|
||||||
deployment:
|
deployment:
|
||||||
storage_secrets: True
|
storage_secrets: true
|
||||||
ceph: True
|
ceph: true
|
||||||
rbd_provisioner: True
|
rbd_provisioner: true
|
||||||
cephfs_provisioner: True
|
cephfs_provisioner: true
|
||||||
client_secrets: False
|
client_secrets: false
|
||||||
rgw_keystone_user_and_endpoints: False
|
rgw_keystone_user_and_endpoints: false
|
||||||
bootstrap:
|
bootstrap:
|
||||||
enabled: True
|
enabled: true
|
||||||
conf:
|
conf:
|
||||||
ceph:
|
ceph:
|
||||||
global:
|
global:
|
||||||
#NOTE (portdirect): any uuidv4 value is acceptable here, though must
|
fsid: ${CEPH_FS_ID}
|
||||||
# be consistent if you wish to redploy whist maintaining data.
|
|
||||||
fsid: C1150D08-707D-4FEB-85E7-692BD9206DC3
|
|
||||||
rgw_ks:
|
rgw_ks:
|
||||||
enabled: True
|
enabled: true
|
||||||
pool:
|
pool:
|
||||||
crush:
|
crush:
|
||||||
#NOTE(portdirect): set to hammer if using kernel < 4.5
|
tunables: ${CRUSH_TUNABLES}
|
||||||
tunables: null
|
|
||||||
target:
|
target:
|
||||||
# NOTE(portdirect): 5 nodes, with one osd per node
|
# NOTE(portdirect): 5 nodes, with one osd per node
|
||||||
osd: 5
|
osd: 5
|
||||||
@ -112,8 +107,7 @@ data:
|
|||||||
type: directory
|
type: directory
|
||||||
location: /var/lib/openstack-helm/ceph/osd/journal-one
|
location: /var/lib/openstack-helm/ceph/osd/journal-one
|
||||||
logging:
|
logging:
|
||||||
fluentd:
|
fluentd: true
|
||||||
enabled: true
|
|
||||||
source:
|
source:
|
||||||
type: local
|
type: local
|
||||||
location: ${OSH_PATH}
|
location: ${OSH_PATH}
|
||||||
@ -146,32 +140,27 @@ data:
|
|||||||
namespace: ceph
|
namespace: ceph
|
||||||
ceph_mon:
|
ceph_mon:
|
||||||
namespace: ceph
|
namespace: ceph
|
||||||
fluentd:
|
|
||||||
namespace: osh-infra
|
|
||||||
network:
|
network:
|
||||||
public: ${CEPH_NETWORK}
|
public: ${CEPH_NETWORK}
|
||||||
cluster: ${CEPH_NETWORK}
|
cluster: ${CEPH_NETWORK}
|
||||||
deployment:
|
deployment:
|
||||||
storage_secrets: True
|
storage_secrets: true
|
||||||
ceph: True
|
ceph: true
|
||||||
rbd_provisioner: True
|
rbd_provisioner: true
|
||||||
cephfs_provisioner: True
|
cephfs_provisioner: true
|
||||||
client_secrets: False
|
client_secrets: false
|
||||||
rgw_keystone_user_and_endpoints: False
|
rgw_keystone_user_and_endpoints: false
|
||||||
bootstrap:
|
bootstrap:
|
||||||
enabled: True
|
enabled: true
|
||||||
conf:
|
conf:
|
||||||
ceph:
|
ceph:
|
||||||
global:
|
global:
|
||||||
#NOTE (portdirect): any uuidv4 value is acceptable here, though must
|
fsid: ${CEPH_FS_ID}
|
||||||
# be consistent if you wish to redploy whist maintaining data.
|
|
||||||
fsid: C1150D08-707D-4FEB-85E7-692BD9206DC3
|
|
||||||
rgw_ks:
|
rgw_ks:
|
||||||
enabled: True
|
enabled: true
|
||||||
pool:
|
pool:
|
||||||
crush:
|
crush:
|
||||||
#NOTE(portdirect): set to hammer if using kernel < 4.5
|
tunables: ${CRUSH_TUNABLES}
|
||||||
tunables: null
|
|
||||||
target:
|
target:
|
||||||
# NOTE(portdirect): 5 nodes, with one osd per node
|
# NOTE(portdirect): 5 nodes, with one osd per node
|
||||||
osd: 5
|
osd: 5
|
||||||
@ -185,8 +174,7 @@ data:
|
|||||||
type: directory
|
type: directory
|
||||||
location: /var/lib/openstack-helm/ceph/osd/journal-one
|
location: /var/lib/openstack-helm/ceph/osd/journal-one
|
||||||
logging:
|
logging:
|
||||||
fluentd:
|
fluentd: true
|
||||||
enabled: true
|
|
||||||
source:
|
source:
|
||||||
type: local
|
type: local
|
||||||
location: ${OSH_PATH}
|
location: ${OSH_PATH}
|
||||||
@ -237,21 +225,23 @@ data:
|
|||||||
public: ${CEPH_NETWORK}
|
public: ${CEPH_NETWORK}
|
||||||
cluster: ${CEPH_NETWORK}
|
cluster: ${CEPH_NETWORK}
|
||||||
deployment:
|
deployment:
|
||||||
storage_secrets: True
|
storage_secrets: true
|
||||||
ceph: True
|
ceph: true
|
||||||
rbd_provisioner: True
|
rbd_provisioner: true
|
||||||
cephfs_provisioner: True
|
cephfs_provisioner: true
|
||||||
client_secrets: False
|
client_secrets: false
|
||||||
rgw_keystone_user_and_endpoints: False
|
rgw_keystone_user_and_endpoints: false
|
||||||
bootstrap:
|
bootstrap:
|
||||||
enabled: True
|
enabled: true
|
||||||
conf:
|
conf:
|
||||||
|
ceph:
|
||||||
|
global:
|
||||||
|
fsid: ${CEPH_FS_ID}
|
||||||
rgw_ks:
|
rgw_ks:
|
||||||
enabled: True
|
enabled: true
|
||||||
pool:
|
pool:
|
||||||
crush:
|
crush:
|
||||||
#NOTE(portdirect): set to hammer if using kernel < 4.5
|
tunables: ${CRUSH_TUNABLES}
|
||||||
tunables: null
|
|
||||||
target:
|
target:
|
||||||
# NOTE(portdirect): 5 nodes, with one osd per node
|
# NOTE(portdirect): 5 nodes, with one osd per node
|
||||||
osd: 5
|
osd: 5
|
||||||
|
@ -132,6 +132,8 @@ data:
|
|||||||
job:
|
job:
|
||||||
node_selector_key: openstack-control-plane
|
node_selector_key: openstack-control-plane
|
||||||
node_selector_value: enabled
|
node_selector_value: enabled
|
||||||
|
bootstrap:
|
||||||
|
enabled: true
|
||||||
source:
|
source:
|
||||||
type: local
|
type: local
|
||||||
location: ${OSH_INFRA_PATH}
|
location: ${OSH_INFRA_PATH}
|
||||||
@ -165,6 +167,8 @@ data:
|
|||||||
auth:
|
auth:
|
||||||
admin:
|
admin:
|
||||||
password: ${DB_ADMIN_PASSWORD}
|
password: ${DB_ADMIN_PASSWORD}
|
||||||
|
volume:
|
||||||
|
enabled: false
|
||||||
labels:
|
labels:
|
||||||
server:
|
server:
|
||||||
node_selector_key: openstack-control-plane
|
node_selector_key: openstack-control-plane
|
||||||
|
@ -93,17 +93,17 @@ data:
|
|||||||
public: ${CEPH_NETWORK}
|
public: ${CEPH_NETWORK}
|
||||||
cluster: ${CEPH_NETWORK}
|
cluster: ${CEPH_NETWORK}
|
||||||
deployment:
|
deployment:
|
||||||
storage_secrets: False
|
storage_secrets: false
|
||||||
ceph: False
|
ceph: false
|
||||||
rbd_provisioner: False
|
rbd_provisioner: false
|
||||||
cephfs_provisioner: False
|
cephfs_provisioner: false
|
||||||
client_secrets: True
|
client_secrets: true
|
||||||
rgw_keystone_user_and_endpoints: False
|
rgw_keystone_user_and_endpoints: false
|
||||||
bootstrap:
|
bootstrap:
|
||||||
enabled: False
|
enabled: false
|
||||||
conf:
|
conf:
|
||||||
rgw_ks:
|
rgw_ks:
|
||||||
enabled: True
|
enabled: true
|
||||||
source:
|
source:
|
||||||
type: local
|
type: local
|
||||||
location: ${OSH_PATH}
|
location: ${OSH_PATH}
|
||||||
@ -718,10 +718,6 @@ data:
|
|||||||
chart_name: libvirt
|
chart_name: libvirt
|
||||||
release: libvirt
|
release: libvirt
|
||||||
namespace: openstack
|
namespace: openstack
|
||||||
wait:
|
|
||||||
timeout: 1800
|
|
||||||
labels:
|
|
||||||
release_group: osh-libvirt
|
|
||||||
install:
|
install:
|
||||||
no_hooks: False
|
no_hooks: False
|
||||||
upgrade:
|
upgrade:
|
||||||
|
Loading…
x
Reference in New Issue
Block a user