From 1258061410908f62c247b437fcb12d2e478ac42d Mon Sep 17 00:00:00 2001
From: OpenStack Proposal Bot <openstack-infra@lists.openstack.org>
Date: Thu, 16 Jan 2020 07:54:06 +0000
Subject: [PATCH] Imported Translations from Zanata

For more information about this automatic import see:
https://docs.openstack.org/i18n/latest/reviewing-translation-import.html

Change-Id: I7e149c96ee4ae090e47f775000dc1e3e55a3efc2
---
 .../locale/en_GB/LC_MESSAGES/doc-testing.po   | 731 ++++++++++++++++++
 1 file changed, 731 insertions(+)
 create mode 100644 doc/source/locale/en_GB/LC_MESSAGES/doc-testing.po

diff --git a/doc/source/locale/en_GB/LC_MESSAGES/doc-testing.po b/doc/source/locale/en_GB/LC_MESSAGES/doc-testing.po
new file mode 100644
index 0000000000..3f264937f9
--- /dev/null
+++ b/doc/source/locale/en_GB/LC_MESSAGES/doc-testing.po
@@ -0,0 +1,731 @@
+# Andi Chandler <andi@gowling.com>, 2019. #zanata
+# Andi Chandler <andi@gowling.com>, 2020. #zanata
+msgid ""
+msgstr ""
+"Project-Id-Version: openstack-helm\n"
+"Report-Msgid-Bugs-To: \n"
+"POT-Creation-Date: 2020-01-13 21:14+0000\n"
+"MIME-Version: 1.0\n"
+"Content-Type: text/plain; charset=UTF-8\n"
+"Content-Transfer-Encoding: 8bit\n"
+"PO-Revision-Date: 2020-01-15 10:36+0000\n"
+"Last-Translator: Andi Chandler <andi@gowling.com>\n"
+"Language-Team: English (United Kingdom)\n"
+"Language: en_GB\n"
+"X-Generator: Zanata 4.3.3\n"
+"Plural-Forms: nplurals=2; plural=(n != 1)\n"
+
+msgid ""
+"1) Initial Ceph and OpenStack deployment: Install Ceph and OpenStack charts "
+"on 3 nodes (mnode1, mnode2 and mnode3). Capture Ceph cluster status as well "
+"as K8s PODs status."
+msgstr ""
+"1) Initial Ceph and OpenStack deployment: Install Ceph and OpenStack charts "
+"on 3 nodes (mnode1, mnode2 and mnode3). Capture Ceph cluster status as well "
+"as K8s PODs status."
+
+msgid ""
+"1) Node reduction: Shutdown 1 of 3 nodes to simulate node failure. Capture "
+"effect of node failure on Ceph as well as other OpenStack services that are "
+"using Ceph."
+msgstr ""
+"1) Node reduction: Shutdown 1 of 3 nodes to simulate node failure. Capture "
+"effect of node failure on Ceph as well as other OpenStack services that are "
+"using Ceph."
+
+msgid "1) Remove out of quorum MON:"
+msgstr "1) Remove out of quorum MON:"
+
+msgid ""
+"2) Node expansion: Apply Ceph and OpenStack related labels to another unused "
+"k8 node. Node expansion should provide more resources for k8 to schedule "
+"PODs for Ceph and OpenStack services."
+msgstr ""
+"2) Node expansion: Apply Ceph and OpenStack related labels to another unused "
+"k8 node. Node expansion should provide more resources for k8 to schedule "
+"PODs for Ceph and OpenStack services."
+
+msgid ""
+"2) Node reduction (failure): Shutdown 1 of 3 nodes (mnode3) to test node "
+"failure. This should cause Ceph cluster to go in HEALTH_WARN state as it has "
+"lost 1 MON and 1 OSD. Capture Ceph cluster status as well as K8s PODs status."
+msgstr ""
+"2) Node reduction (failure): Shutdown 1 of 3 nodes (mnode3) to test node "
+"failure. This should cause Ceph cluster to go in HEALTH_WARN state as it has "
+"lost 1 MON and 1 OSD. Capture Ceph cluster status as well as K8s PODs status."
+
+msgid "2) Remove down OSD from Ceph cluster:"
+msgstr "2) Remove down OSD from Ceph cluster:"
+
+msgid ""
+"3) Fix Ceph Cluster: After node expansion, perform maintenance on Ceph "
+"cluster to ensure quorum is reached and Ceph is HEALTH_OK."
+msgstr ""
+"3) Fix Ceph Cluster: After node expansion, perform maintenance on Ceph "
+"cluster to ensure quorum is reached and Ceph is HEALTH_OK."
+
+msgid ""
+"3) Node expansion: Add Ceph and OpenStack related labels to 4th node "
+"(mnode4) for expansion. Ceph cluster would show new MON and OSD being added "
+"to cluster. However Ceph cluster would continue to show HEALTH_WARN because "
+"1 MON and 1 OSD are still missing."
+msgstr ""
+"3) Node expansion: Add Ceph and OpenStack related labels to 4th node "
+"(mnode4) for expansion. Ceph cluster would show new MON and OSD being added "
+"to cluster. However Ceph cluster would continue to show HEALTH_WARN because "
+"1 MON and 1 OSD are still missing."
+
+msgid ""
+"4) Ceph cluster recovery: Perform Ceph maintenance to make Ceph cluster "
+"HEALTH_OK. Remove lost MON and OSD from Ceph cluster."
+msgstr ""
+"4) Ceph cluster recovery: Perform Ceph maintenance to make Ceph cluster "
+"HEALTH_OK. Remove lost MON and OSD from Ceph cluster."
+
+msgid ""
+"4. Replace the failed disk with a new one. If you repair (not replace) the "
+"failed disk, you may need to run the following:"
+msgstr ""
+"4. Replace the failed disk with a new one. If you repair (not replace) the "
+"failed disk, you may need to run the following:"
+
+msgid "6 Nodes (VM based) env"
+msgstr "6 Nodes (VM based) env"
+
+msgid ""
+"Above output shows Ceph cluster in HEALTH_OK with all OSDs and MONs up and "
+"running."
+msgstr ""
+"Above output shows Ceph cluster in HEALTH_OK with all OSDs and MONs up and "
+"running."
+
+msgid "Above output shows that ``osd.1`` is down."
+msgstr "Above output shows that ``osd.1`` is down."
+
+msgid "Adding Tests"
+msgstr "Adding Tests"
+
+msgid ""
+"Additional information on Helm tests for OpenStack-Helm and how to execute "
+"these tests locally via the scripts used in the gate can be found in the "
+"gates_ directory."
+msgstr ""
+"Additional information on Helm tests for OpenStack-Helm and how to execute "
+"these tests locally via the scripts used in the gate can be found in the "
+"gates_ directory."
+
+msgid ""
+"After 10+ miniutes, Ceph starts rebalancing with one node lost (i.e., 6 osds "
+"down) and the status stablizes with 18 osds."
+msgstr ""
+"After 10+ miniutes, Ceph starts rebalancing with one node lost (i.e., 6 osds "
+"down) and the status stablises with 18 osds."
+
+msgid "After applying labels, let's check status"
+msgstr "After applying labels, let's check status"
+
+msgid "After reboot (node voyager3), the node status changes to ``NotReady``."
+msgstr "After reboot (node voyager3), the node status changes to ``NotReady``."
+
+msgid "All PODs are in running state."
+msgstr "All PODs are in running state."
+
+msgid ""
+"All tests should be added to the gates during development, and are required "
+"for any new service charts prior to merging.  All Helm tests should be "
+"included as part of the deployment script.  An example of this can be seen "
+"in this script_."
+msgstr ""
+"All tests should be added to the gates during development, and are required "
+"for any new service charts prior to merging.  All Helm tests should be "
+"included as part of the deployment script.  An example of this can be seen "
+"in this script_."
+
+msgid "Any Helm tests associated with a chart can be run by executing:"
+msgstr "Any Helm tests associated with a chart can be run by executing:"
+
+msgid ""
+"Any templates for Helm tests submitted should follow the philosophies "
+"applied in the other templates.  These include: use of overrides where "
+"appropriate, use of endpoint lookups and other common functionality in helm-"
+"toolkit, and mounting any required scripting templates via the configmap-bin "
+"template for the service chart.  If Rally tests are not appropriate or "
+"adequate for a service chart, any additional tests should be documented "
+"appropriately and adhere to the same expectations."
+msgstr ""
+"Any templates for Helm tests submitted should follow the philosophies "
+"applied in the other templates.  These include: use of overrides where "
+"appropriate, use of endpoint lookups and other common functionality in helm-"
+"toolkit, and mounting any required scripting templates via the configmap-bin "
+"template for the service chart.  If Rally tests are not appropriate or "
+"adequate for a service chart, any additional tests should be documented "
+"appropriately and adhere to the same expectations."
+
+msgid ""
+"As shown above, Ceph status is now HEALTH_OK and shows 3 MONs available."
+msgstr ""
+"As shown above, Ceph status is now HEALTH_OK and shows 3 MONs available."
+
+msgid ""
+"As shown in Ceph status above, ``osd: 4 osds: 3 up, 3 in`` 1 of 4 OSDs is "
+"still down. Let's remove that OSD."
+msgstr ""
+"As shown in Ceph status above, ``osd: 4 osds: 3 up, 3 in`` 1 of 4 OSDs is "
+"still down. Let's remove that OSD."
+
+msgid "Capture Ceph pods statuses."
+msgstr "Capture Ceph pods statuses."
+
+msgid "Capture Openstack pods statuses."
+msgstr "Capture Openstack pods statuses."
+
+msgid "Capture final Ceph pod statuses:"
+msgstr "Capture final Ceph pod statuses:"
+
+msgid "Capture final Openstack pod statuses:"
+msgstr "Capture final Openstack pod statuses:"
+
+msgid "Case: A disk fails"
+msgstr "Case: A disk fails"
+
+msgid "Case: One host machine where ceph-mon is running is rebooted"
+msgstr "Case: One host machine where ceph-mon is running is rebooted"
+
+msgid "Caveats:"
+msgstr "Caveats:"
+
+msgid "Ceph - Node Reduction, Expansion and Ceph Recovery"
+msgstr "Ceph - Node Reduction, Expansion and Ceph Recovery"
+
+msgid "Ceph Cephfs provisioner docker images."
+msgstr "Ceph Cephfs provisioner docker images."
+
+msgid "Ceph Luminous point release images for Ceph components"
+msgstr "Ceph Luminous point release images for Ceph components"
+
+msgid "Ceph MON and OSD PODs got scheduled on mnode4 node."
+msgstr "Ceph MON and OSD PODs got scheduled on mnode4 node."
+
+msgid "Ceph RBD provisioner docker images."
+msgstr "Ceph RBD provisioner docker images."
+
+msgid ""
+"Ceph can be upgraded without downtime for Openstack components in a "
+"multinode env."
+msgstr ""
+"Ceph can be upgraded without downtime for Openstack components in a "
+"multinode env."
+
+msgid "Ceph cluster is in HEALTH_OK state with 3 MONs and 3 OSDs."
+msgstr "Ceph cluster is in HEALTH_OK state with 3 MONs and 3 OSDs."
+
+msgid "Ceph status shows 1 Ceph MON and 1 Ceph OSD missing."
+msgstr "Ceph status shows 1 Ceph MON and 1 Ceph OSD missing."
+
+msgid "Ceph status shows HEALTH_WARN as expected"
+msgstr "Ceph status shows HEALTH_WARN as expected"
+
+msgid "Ceph status shows that MON and OSD count has been increased."
+msgstr "Ceph status shows that MON and OSD count has been increased."
+
+msgid ""
+"Ceph status shows that ceph-mon running on ``voyager3`` becomes out of "
+"quorum. Also, six osds running on ``voyager3`` are down; i.e., 18 osds are "
+"up out of 24 osds."
+msgstr ""
+"Ceph status shows that ceph-mon running on ``voyager3`` becomes out of "
+"quorum. Also, six osds running on ``voyager3`` are down; i.e., 18 osds are "
+"up out of 24 osds."
+
+msgid "Ceph status still shows HEALTH_WARN as one MON and OSD are still down."
+msgstr "Ceph status still shows HEALTH_WARN as one MON and OSD are still down."
+
+msgid "Ceph version: 12.2.3"
+msgstr "Ceph version: 12.2.3"
+
+msgid "Check Ceph Pods"
+msgstr "Check Ceph Pods"
+
+msgid "Check version of each Ceph components."
+msgstr "Check version of each Ceph components."
+
+msgid "Check which images Provisionors and Mon-Check PODs are using"
+msgstr "Check which images Provisionors and Mon-Check PODs are using"
+
+msgid "Cluster size: 4 host machines"
+msgstr "Cluster size: 4 host machines"
+
+msgid "Conclusion:"
+msgstr "Conclusion:"
+
+msgid "Confirm Ceph component's version."
+msgstr "Confirm Ceph component's version."
+
+msgid "Continue with OSH multinode guide to install other Openstack charts."
+msgstr "Continue with OSH multinode guide to install other Openstack charts."
+
+msgid "Deploy and Validate Ceph"
+msgstr "Deploy and Validate Ceph"
+
+msgid "Disk Failure"
+msgstr "Disk Failure"
+
+msgid "Docker Images:"
+msgstr "Docker Images:"
+
+msgid ""
+"Every OpenStack-Helm chart should include any required Helm tests necessary "
+"to provide a sanity check for the OpenStack service.  Information on using "
+"the Helm testing framework can be found in the Helm repository_.  Currently, "
+"the Rally testing framework is used to provide these checks for the core "
+"services.  The Keystone Helm test template can be used as a reference, and "
+"can be found here_."
+msgstr ""
+"Every OpenStack-Helm chart should include any required Helm tests necessary "
+"to provide a sanity check for the OpenStack service.  Information on using "
+"the Helm testing framework can be found in the Helm repository_.  Currently, "
+"the Rally testing framework is used to provide these checks for the core "
+"services.  The Keystone Helm test template can be used as a reference, and "
+"can be found here_."
+
+msgid "Find that Ceph is healthy with a lost OSD (i.e., a total of 23 OSDs):"
+msgstr "Find that Ceph is healthy with a lost OSD (i.e., a total of 23 OSDs):"
+
+msgid "First, run ``ceph osd tree`` command to get list of OSDs."
+msgstr "First, run ``ceph osd tree`` command to get list of OSDs."
+
+msgid "Follow all steps from OSH multinode guide with below changes."
+msgstr "Follow all steps from OSH multinode guide with below changes."
+
+msgid ""
+"Followed OSH multinode guide steps to install Ceph and OpenStack charts up "
+"to Cinder."
+msgstr ""
+"Followed OSH multinode guide steps to install Ceph and OpenStack charts up "
+"to Cinder."
+
+msgid ""
+"Followed OSH multinode guide steps to setup nodes and install K8s cluster"
+msgstr ""
+"Followed OSH multinode guide steps to setup nodes and install K8s cluster"
+
+msgid "Following is a partial part from script to show changes."
+msgstr "Following is a partial part from script to show changes."
+
+msgid ""
+"From the Kubernetes cluster, remove the failed OSD pod, which is running on "
+"``voyager4``:"
+msgstr ""
+"From the Kubernetes cluster, remove the failed OSD pod, which is running on "
+"``voyager4``:"
+
+msgid "Hardware Failure"
+msgstr "Hardware Failure"
+
+msgid "Helm Tests"
+msgstr "Helm Tests"
+
+msgid "Host Failure"
+msgstr "Host Failure"
+
+msgid ""
+"In this test env, MariaDB chart is deployed with only 1 replica. In order to "
+"test properly, the node with MariaDB server POD (mnode2) should not be "
+"shutdown."
+msgstr ""
+"In this test env, MariaDB chart is deployed with only 1 replica. In order to "
+"test properly, the node with MariaDB server POD (mnode2) should not be "
+"shutdown."
+
+msgid "In this test env, ``mnode3`` is out of quorum."
+msgstr "In this test env, ``mnode3`` is out of quorum."
+
+msgid ""
+"In this test env, each node has Ceph and OpenStack related PODs. Due to "
+"this, shutting down a Node will cause issue with Ceph as well as OpenStack "
+"services. These PODs level failures are captured following subsequent "
+"screenshots."
+msgstr ""
+"In this test env, each node has Ceph and OpenStack related PODs. Due to "
+"this, shutting down a Node will cause issue with Ceph as well as OpenStack "
+"services. These PODs level failures are captured following subsequent "
+"screenshots."
+
+msgid "In this test env, let's shutdown ``mnode3`` node."
+msgstr "In this test env, let's shutdown ``mnode3`` node."
+
+msgid ""
+"In this test env, let's use ``mnode4`` and apply Ceph and OpenStack related "
+"labels."
+msgstr ""
+"In this test env, let's use ``mnode4`` and apply Ceph and OpenStack related "
+"labels."
+
+msgid ""
+"In this test env, since out of quorum MON is no longer available due to node "
+"failure, we can processed with removing it from Ceph cluster."
+msgstr ""
+"In this test env, since out of quorum MON is no longer available due to node "
+"failure, we can processed with removing it from Ceph cluster."
+
+msgid "Install Ceph charts (version 12.2.4)"
+msgstr "Install Ceph charts (version 12.2.4)"
+
+msgid "Install OSH components as per OSH multinode guide."
+msgstr "Install OSH components as per OSH multinode guide."
+
+msgid "Install Openstack charts"
+msgstr "Install Openstack charts"
+
+msgid "Kubernetes version: 1.10.5"
+msgstr "Kubernetes version: 1.10.5"
+
+msgid "Let's add more resources for K8s to schedule PODs on."
+msgstr "Let's add more resources for K8s to schedule PODs on."
+
+msgid ""
+"Make sure only 3 nodes (mnode1, mnode2, mnode3) have Ceph and OpenStack "
+"related labels. K8s would only schedule PODs on these 3 nodes."
+msgstr ""
+"Make sure only 3 nodes (mnode1, mnode2, mnode3) have Ceph and OpenStack "
+"related labels. K8s would only schedule PODs on these 3 nodes."
+
+msgid "Mission"
+msgstr "Mission"
+
+msgid ""
+"Note: To find the daemonset associated with a failed OSD, check out the "
+"followings:"
+msgstr ""
+"Note: To find the daemonset associated with a failed OSD, check out the "
+"followings:"
+
+msgid ""
+"Now that we have added new node for Ceph and OpenStack PODs, let's perform "
+"maintenance on Ceph cluster."
+msgstr ""
+"Now that we have added new node for Ceph and OpenStack PODs, let's perform "
+"maintenance on Ceph cluster."
+
+msgid "Number of disks: 24 (= 6 disks per host * 4 hosts)"
+msgstr "Number of disks: 24 (= 6 disks per host * 4 hosts)"
+
+msgid "OSD count is set to 3 based on env setup."
+msgstr "OSD count is set to 3 based on env setup."
+
+msgid ""
+"Only 3 nodes will have Ceph and OpenStack related labels. Each of these 3 "
+"nodes will have one MON and one OSD running on them."
+msgstr ""
+"Only 3 nodes will have Ceph and OpenStack related labels. Each of these 3 "
+"nodes will have one MON and one OSD running on them."
+
+msgid "OpenStack PODs that were scheduled mnode3 also shows NodeLost/Unknown."
+msgstr "OpenStack PODs that were scheduled mnode3 also shows NodeLost/Unknown."
+
+msgid "OpenStack-Helm commit: 25e50a34c66d5db7604746f4d2e12acbdd6c1459"
+msgstr "OpenStack-Helm commit: 25e50a34c66d5db7604746f4d2e12acbdd6c1459"
+
+msgid ""
+"Our focus lies on resiliency for various failure scenarios but not on "
+"performance or stress testing."
+msgstr ""
+"Our focus lies on resiliency for various failure scenarios but not on "
+"performance or stress testing."
+
+msgid "PODs that were scheduled on mnode3 node has status of NodeLost/Unknown."
+msgstr ""
+"PODs that were scheduled on mnode3 node has status of NodeLost/Unknown."
+
+msgid "Recovery:"
+msgstr "Recovery:"
+
+msgid ""
+"Remove the failed OSD (OSD ID = 2 in this example) from the Ceph cluster:"
+msgstr ""
+"Remove the failed OSD (OSD ID = 2 in this example) from the Ceph cluster:"
+
+msgid "Resiliency Tests for OpenStack-Helm/Ceph"
+msgstr "Resiliency Tests for OpenStack-Helm/Ceph"
+
+msgid "Run ``ceph osd purge`` command to remove OSD from ceph cluster."
+msgstr "Run ``ceph osd purge`` command to remove OSD from Ceph cluster."
+
+msgid "Running Tests"
+msgstr "Running Tests"
+
+msgid "Setup:"
+msgstr "Setup:"
+
+msgid ""
+"Showing partial output from kubectl describe command to show which image is "
+"Docker container is using"
+msgstr ""
+"Showing partial output from kubectl describe command to show which image is "
+"Docker container is using"
+
+msgid ""
+"Shutdown 1 of 3 nodes (mnode1, mnode2, mnode3) to simulate node failure/lost."
+msgstr ""
+"Shutdown 1 of 3 nodes (mnode1, mnode2, mnode3) to simulate node failure/lost."
+
+msgid ""
+"Since the node that was shutdown earlier had both Ceph and OpenStack PODs, "
+"mnode4 should get Ceph and OpenStack related labels as well."
+msgstr ""
+"Since the node that was shutdown earlier had both Ceph and OpenStack PODs, "
+"mnode4 should get Ceph and OpenStack related labels as well."
+
+msgid "Software Failure"
+msgstr "Software Failure"
+
+msgid "Solution:"
+msgstr "Solution:"
+
+msgid "Start a new OSD pod on ``voyager4``:"
+msgstr "Start a new OSD pod on ``voyager4``:"
+
+msgid "Step 1: Initial Ceph and OpenStack deployment"
+msgstr "Step 1: Initial Ceph and OpenStack deployment"
+
+msgid "Step 2: Node reduction (failure):"
+msgstr "Step 2: Node reduction (failure):"
+
+msgid "Step 3: Node Expansion"
+msgstr "Step 3: Node Expansion"
+
+msgid "Step 4: Ceph cluster recovery"
+msgstr "Step 4: Ceph cluster recovery"
+
+msgid "Steps:"
+msgstr "Steps:"
+
+msgid "Symptom:"
+msgstr "Symptom:"
+
+msgid "Test Environment"
+msgstr "Test Environment"
+
+msgid "Test Scenarios:"
+msgstr "Test Scenarios:"
+
+msgid "Testing"
+msgstr "Testing"
+
+msgid "Testing Expectations"
+msgstr "Testing Expectations"
+
+msgid ""
+"The goal of our resiliency tests for `OpenStack-Helm/Ceph <https://github."
+"com/openstack/openstack-helm/tree/master/ceph>`_ is to show symptoms of "
+"software/hardware failure and provide the solutions."
+msgstr ""
+"The goal of our resiliency tests for `OpenStack-Helm/Ceph <https://github."
+"com/openstack/openstack-helm/tree/master/ceph>`_ is to show symptoms of "
+"software/hardware failure and provide the solutions."
+
+msgid ""
+"The node status of ``voyager3`` changes to ``Ready`` after the node is up "
+"again. Also, Ceph pods are restarted automatically. The Ceph status shows "
+"that the monitor running on ``voyager3`` is now in quorum and 6 osds gets "
+"back up (i.e., a total of 24 osds are up)."
+msgstr ""
+"The node status of ``voyager3`` changes to ``Ready`` after the node is up "
+"again. Also, Ceph pods are restarted automatically. The Ceph status shows "
+"that the monitor running on ``voyager3`` is now in quorum and 6 osds gets "
+"back up (i.e., a total of 24 osds are up)."
+
+msgid ""
+"The output of the Helm tests can be seen by looking at the logs of the pod "
+"created by the Helm tests.  These logs can be viewed with:"
+msgstr ""
+"The output of the Helm tests can be seen by looking at the logs of the pod "
+"created by the Helm tests.  These logs can be viewed with:"
+
+msgid "The pod status of ceph-mon and ceph-osd shows as ``NodeLost``."
+msgstr "The pod status of ceph-mon and ceph-osd shows as ``NodeLost``."
+
+msgid ""
+"This document captures steps and result from node reduction and expansion as "
+"well as ceph recovery."
+msgstr ""
+"This document captures steps and result from node reduction and expansion as "
+"well as Ceph recovery."
+
+msgid ""
+"This is to test a scenario when a disk failure happens. We monitor the ceph "
+"status and notice one OSD (osd.2) on voyager4 which has ``/dev/sdh`` as a "
+"backend is down."
+msgstr ""
+"This is to test a scenario when a disk failure happens. We monitor the ceph "
+"status and notice one OSD (osd.2) on voyager4 which has ``/dev/sdh`` as a "
+"backend is down."
+
+msgid "To replace the failed OSD, execute the following procedure:"
+msgstr "To replace the failed OSD, execute the following procedure:"
+
+msgid "Update Ceph Client chart with new overrides:"
+msgstr "Update Ceph Client chart with new overrides:"
+
+msgid "Update Ceph Mon chart with new overrides"
+msgstr "Update Ceph Mon chart with new overrides"
+
+msgid "Update Ceph OSD chart with new overrides:"
+msgstr "Update Ceph OSD chart with new overrides:"
+
+msgid "Update Ceph Provisioners chart with new overrides:"
+msgstr "Update Ceph Provisioners chart with new overrides:"
+
+msgid ""
+"Update ceph install script ``./tools/deployment/multinode/030-ceph.sh`` to "
+"add ``images:`` section in overrides as shown below."
+msgstr ""
+"Update Ceph install script ``./tools/deployment/multinode/030-ceph.sh`` to "
+"add ``images:`` section in overrides as shown below."
+
+msgid ""
+"Update, image section in new overrides ``ceph-update.yaml`` as shown below"
+msgstr ""
+"Update, image section in new overrides ``ceph-update.yaml`` as shown below"
+
+msgid "Upgrade Ceph charts to update version"
+msgstr "Upgrade Ceph charts to update version"
+
+msgid ""
+"Upgrade Ceph charts to version 12.2.5 by updating docker images in overrides."
+msgstr ""
+"Upgrade Ceph charts to version 12.2.5 by updating Docker images in overrides."
+
+msgid ""
+"Use Ceph override file ``ceph.yaml`` that was generated previously and "
+"update images section as below"
+msgstr ""
+"Use Ceph override file ``ceph.yaml`` that was generated previously and "
+"update images section as below"
+
+msgid ""
+"Using ``ceph mon_status`` and ``ceph -s`` commands, confirm ID of MON that "
+"is out of quorum."
+msgstr ""
+"Using ``ceph mon_status`` and ``ceph -s`` commands, confirm ID of MON that "
+"is out of quorum."
+
+msgid ""
+"Validate the Ceph status (i.e., one OSD is added, so the total number of "
+"OSDs becomes 24):"
+msgstr ""
+"Validate the Ceph status (i.e., one OSD is added, so the total number of "
+"OSDs becomes 24):"
+
+msgid "`Disk failure <./disk-failure.html>`_"
+msgstr "`Disk failure <./disk-failure.html>`_"
+
+msgid "`Host failure <./host-failure.html>`_"
+msgstr "`Host failure <./host-failure.html>`_"
+
+msgid "`Monitor failure <./monitor-failure.html>`_"
+msgstr "`Monitor failure <./monitor-failure.html>`_"
+
+msgid "`OSD failure <./osd-failure.html>`_"
+msgstr "`OSD failure <./osd-failure.html>`_"
+
+msgid "``Ceph MON Status:``"
+msgstr "``Ceph MON Status:``"
+
+msgid "``Ceph MON Status``"
+msgstr "``Ceph MON Status``"
+
+msgid "``Ceph PODs:``"
+msgstr "``Ceph PODs:``"
+
+msgid "``Ceph PODs``"
+msgstr "``Ceph PODs``"
+
+msgid "``Ceph Status:``"
+msgstr "``Ceph Status:``"
+
+msgid "``Ceph quorum status:``"
+msgstr "``Ceph quorum status:``"
+
+msgid "``Ceph quorum status``"
+msgstr "``Ceph quorum status``"
+
+msgid "``Ceph status:``"
+msgstr "``Ceph status:``"
+
+msgid "``Ceph status``"
+msgstr "``Ceph status``"
+
+msgid "``Check node status:``"
+msgstr "``Check node status:``"
+
+msgid "``Following are PODs scheduled on mnode3 before shutdown:``"
+msgstr "``Following are PODs scheduled on mnode3 before shutdown:``"
+
+msgid "``OpenStack PODs:``"
+msgstr "``OpenStack PODs:``"
+
+msgid "``OpenStack PODs``"
+msgstr "``OpenStack PODs``"
+
+msgid "``Remove MON from Ceph cluster``"
+msgstr "``Remove MON from Ceph cluster``"
+
+msgid "``Result/Observation:``"
+msgstr "``Result/Observation:``"
+
+msgid ""
+"``Results:`` All provisioner pods got terminated at once (same time). Other "
+"ceph pods are running. No interruption to OSH pods."
+msgstr ""
+"``Results:`` All provisioner pods got terminated at once (same time). Other "
+"Ceph pods are running. No interruption to OSH pods."
+
+msgid ""
+"``Results:`` Mon pods got updated one by one (rolling updates). Each Mon pod "
+"got respawn and was in 1/1 running state before next Mon pod got updated. "
+"Each Mon pod got restarted. Other ceph pods were not affected with this "
+"update. No interruption to OSH pods."
+msgstr ""
+"``Results:`` Mon pods got updated one by one (rolling updates). Each Mon pod "
+"got respawn and was in 1/1 running state before next Mon pod got updated. "
+"Each Mon pod got restarted. Other ceph pods were not affected with this "
+"update. No interruption to OSH pods."
+
+msgid ""
+"``Results:`` Rolling updates (one pod at a time). Other ceph pods are "
+"running. No interruption to OSH pods."
+msgstr ""
+"``Results:`` Rolling updates (one pod at a time). Other ceph pods are "
+"running. No interruption to OSH pods."
+
+msgid ""
+"``ceph_bootstrap``, ``ceph-config_helper`` and ``ceph_rbs_pool`` images are "
+"used for jobs. ``ceph_mon_check`` has one script that is stable so no need "
+"to upgrade."
+msgstr ""
+"``ceph_bootstrap``, ``ceph-config_helper`` and ``ceph_rbs_pool`` images are "
+"used for jobs. ``ceph_mon_check`` has one script that is stable so no need "
+"to upgrade."
+
+msgid "``cp /tmp/ceph.yaml ceph-update.yaml``"
+msgstr "``cp /tmp/ceph.yaml ceph-update.yaml``"
+
+msgid "``helm upgrade ceph-client ./ceph-client --values=ceph-update.yaml``"
+msgstr "``helm upgrade ceph-client ./ceph-client --values=ceph-update.yaml``"
+
+msgid "``helm upgrade ceph-mon ./ceph-mon --values=ceph-update.yaml``"
+msgstr "``helm upgrade ceph-mon ./ceph-mon --values=ceph-update.yaml``"
+
+msgid "``helm upgrade ceph-osd ./ceph-osd --values=ceph-update.yaml``"
+msgstr "``helm upgrade ceph-osd ./ceph-osd --values=ceph-update.yaml``"
+
+msgid ""
+"``helm upgrade ceph-provisioners ./ceph-provisioners --values=ceph-update."
+"yaml``"
+msgstr ""
+"``helm upgrade ceph-provisioners ./ceph-provisioners --values=ceph-update."
+"yaml``"
+
+msgid "``series of console outputs:``"
+msgstr "``series of console outputs:``"