Management port exposes a HTTP REST API which can be used to
add virtual hosts and users, gather statistics and other data
about the running cluster. This also turns on the web interface
available on the management port.
Change-Id: I5e4c902d780d99e96fdd160c58227676d4676e87
Partial-Implements: blueprint service-rabbitmq-registration
This PS removes the licence header from rendered output from tiller,
significantly reducing the configmap size of charts deployed to the
cluster.
Change-Id: I5d1b246f2068f3b83bf59ba79fe8b88bbc9a6161
This PS move s the replicas key to be under the pod key in the values.
It brings further consolation of related configuration params to be
nested under common keys across all charts.
Change-Id: I420b06debd0a62ba5d83497be43ff6c49c49d339
This patchset enforces stricter file permission on *-etc configmap and
sets readOnly flag to true in a number of charts.
Change-Id: I233689a5d56dd1352e0d81997a94b4cdd6bed5d2
Signed-off-by: Tin Lam <tin@irrational.io>
This PS adds soft anti-affinity to all pods in OS-H. By doing so
resiliancy is improved by attempting to ensure that pods are created
on seperate nodes.
Change-Id: I0c1092498f7a1e44218ef785ca3f73fa9f49819c
K8s 1.6 has affinity/antiaffinity funtionalities as a beta feature. This
means we don't need to declare them using annotations. This commit
switches usages of affinity to 1.6 syntax.
Change-Id: Ia68f4ab28a018617bd44b1295fea58cd30eb4a39
With 1.6, init containers are officially part of the kubernetes
API. This changes the format of the helm template for the
entrypoint container from json to yaml, and updates the
charts accordingly.
Co-Authored-By: Pete Birley <pete@port.direct>
Change-Id: I569566ce4b031d107af2d38483040a26210bec45
This PS updates the way helm-toolkit functions are named to
reference the full path they are loacted at. This should make
development and debugging easier. Addtionally unused functions
have been pruned as well.
Change-Id: I03c553f1d01bccc70c86768b416b147c90d9b2f0
This change removes any trailing whitespaces found while
running:
find . -type f -exec egrep -l " +$" {} \;
Change-Id: Ic1dc8765a0971a32da4c3116871e9450a55b357f
Currently, rabbitmq clustering is using the autocluster plugin,
and NODE_TYPE is set to disc in default. so every nodes join cluster as disc node.
However, since there is a need for disc + ram clustering for the performance.
if change node_type from disc to ram, clustering configured as [disc + ram + ram].
Change-Id: Ie83689b0554f0f993bdffac666f0f56db8082992
Implements: blueprint rabbitmq-dns-discovery
Some useful things to note:
1. This uses a StatefulSet instead of a Deployment. The reason for this
is that when RabbitMQ uses DNS for peer discovery, the first thing it
does when trying to join a node is attempt a reverse-dns lookup.
This reverse lookup works when using a StatefulSet, but not a
Deployment.
2. The RabbitMQ configuration was updated to use the new sysctl-style
format. It seems that the new format is required to configure the
new autoclustering features. Additionally, I found that this
generate much clearer error messages than the straight erlang format.
3. I removed the `is-node-properly-clustered` test in the liveness and
readiness probes. This probe isn't directly supported in 3.7.0,
and it wasn't clear that a clustering check was appropriate for each
node.
Change-Id: Ieefbb2205bd77fbac04abcd051fb06fce62e8d97
* Rename common chart to helm-toolkit
* Update useage of helpers to include reference to chart they come from.
* Update helm-toolkit function naming
Also catches several functions missed in previous PS
* Update remaining requirements.yaml to use helm-toolbox
* Dep Check container fix for common -> helm-toolbox renaming
* Add resources limits and requests for each chart
* Refactor the resources limits and requests to follow a patern
* Fix some coding issues
* fix issues resulting from feedback on the resources PR
* Reset some variables to a static value in the neutron chart.
* Substituting variable entrypoint by dependency_check in the concerned files
* Few adjustments
* Update deploy-region.yaml
* Update deployment.yaml
* Add resources limits and requests for each chart
Squah all commits in one.
* Add resources limits and requests for some charts
* cleaning
* Fix indendation issue
* Update deployment.yaml
* Update daemonset-ovs-vswitchd.yaml
* These annotations ensure that if configmaps change, the pods
are redeployed according to their upgrade strategy.
* This change excludes glance which has a pull request in progress
to refactor it.
* As of helm 2.0.0, it no longer seems necessary to
copy a "globals.tpl" into each child chart, defines in
child charts now seem to be reliably rendered in the
parent. This was not working correctly in 2.0.0rc2
and so after a lot of testing, the globals define
insertion has been removed from the Makefiles. Also,
includes piped to b64encode are working now, meaning
we no longer have to do Makefile magic to base64
encode successfully. That has been removed. Once
.Files.Get works properly in a child chart context
we can remove all Make help.
* The openstack-base chart has been renamed to common
to better reflect that it is common to everything in
aic-helm, even non-openstack projects like ceph. All
charts now include it as a requirement.
* A first pass at a "cleaner" template directory approach
has been applied to rabbitmq as a test chart. This allows
files such as _start_rabbit to live in their raw form
and organizes files by config vs script in 'etc' or 'bin'
directories. If accepted, will apply to all other charts
for consistency.
* Made values.yaml consistent throughout charts. Removed any globals
references in subcharts as these are difficult to override. Only
ports should be in globals to build URLs which can come as part
of a future commit. The hostname endpoint aspect of a service
will come from openstack-base/_hosts.tpl and the port
would come from the chart itself as a global so other charts
can reference the port to build a complete URL. Putting the
hostnames themselves as globals in individual charts makes it
difficult to make a sweeping top level FQDN change.
* Cleaned up yaml requirements and incorporated a new _common.tpl
that is distributed to all charts to allow common endpoint naming
while still retaining the ability to install individual charts.
* Fixed keystone URL generation during bootstrap as a correct
URL is critical given keystone uses this to construct all
subsequent URLs in the request. Also allow controlling the
default endpoint version and scheme.
* Added missing NAMESPACE declaration to keystone deployment
as this is required for entrypoint to discover resources
not in the 'default' namespace.
* Refactored all nodeSelector values to be consistent throughout
all charts