This is to export the ceph cluster name as environment variable
since its getting referred by scripts.
also to fix the query to get inactive pgs.
Change-Id: I1db5cfbd594c0cc6d54f748f22af5856d9594922
The PS updates queries in wait_for_pgs function in ceph-client and
ceph-osd charts. It allows more accurately check the status of PGs.
The output of the "ceph pg ls" command may contain many PG statuses,
like "active+clean", "active+undersized+degraded", "active+recovering",
"peering" and etc. But along with these statuses there may be such as
"stale+active+clean". To avoid the wrong interpretation of the status
of the PSs the filter was changed from "startswith(active+)" to
"contains(active)".
Also PS adds a delay after restart of the pods to post-apply job.
It allows to reduce the number of useless queries to kubernetes.
Change-Id: I0eff2ce036ad543bf2554bd586c2a2d3e91c052b
Adds configuration options for the --default-ssl-certificate feature of
NGINX Ingress Controller, which provides a default certificate for
requests that do not match any configured server names.[0]
To enable with a new certificate, specify:
.conf.default_ssl_certificate.enabled=true
.endpoints.ingress.host_fqdn_override.public.tls.crt="PEM cert data"
.endpoints.ingress.host_fqdn_override.public.tls.key="PEM key data"
.manifests.secret_ingress_tls=true
To enable using a TLS cert in an existing secret, specify:
.conf.default_ssl_certificate.enabled=true
.conf.default_ssl_certificate.name="name of the secret"
.conf.default_ssl_certificate.namespace="namespace of the secret"
0: https://kubernetes.github.io/ingress-nginx/user-guide/tls/#default-ssl-certificate
Change-Id: Idd704fd880f56137923d4c38cc188b130ee3b56d
The PS switches back to old naming convention for VGs. The old naming
convention have to be used until the changes which allow to handle
update of VG names are merged ( https://review.opendev.org/#/c/745166/ ).
Otherwise, OSDs will not come up after an upgrade.
Change-Id: I1bf9ca93149a93dfd5f79813533ace3a1fe58002
If the test pod still exists, then the new test run
fails with ERROR: pods "abc-test" already exists
So, Removing remaining test pods before new test run
Change-Id: I3b3ed5ceaf420aa39a669b4a50a838ad154b1fdd
Closes-Bug: #1882030
The recently-added crush weight comparison in reweight_osds() that
checks weights for zero isn't working correctly because the
expected weight is being calculated to two decimal places and then
compared against "0" as a string. This updates the comparison
string to "0.00" to match the calculation.
Change-Id: I29387a597a21180bb7fba974b4daeadf6ffc182d
The PS updates post apply job and moves execution of the command outside
of if statement. The output of the command stored in a variable
which will be checked in if statement. Added "-z" to correct comparison
of the length of the string (variable). It was accidentally missed in
the initial PS.
Change-Id: I907f75d0a9e5ef27fba5306ddb86199e94b01b3b
1) A separate prometheus job need to provide target and scrap metrics
2) it is based on https://github.com/prometheus/blackbox_exporter
Adding script file for deployment and job under zuul.d
Resolving conflict
Change-Id: Ia15ab7d8ef882886fe0e37cc2599e6815d7bcc6c
If circumstances are such that the reweight function believes
OSD disks have zero size, refrain from reweighting OSDs to 0.
This can happen if OSDs are deployed with the noup flag set.
Also move the setting and unsetting of flags above this
calculation as an additional precautionary measure.
Change-Id: Ibc23494e0e75cfdd7654f5c0d3b6048b146280f7
This implements security context override at pod level and adds
readOnly-fs to keystone-webhook container
Change-Id: Ia67947b7323e41363a5ee379c0dfb001936b5107
During bootstrap process kubernetes node is not ready due to missed CNI.
It will be installed later but for a few deployments/jobs it's critical.
They can't start pods and looping in a while.
Workaround is here: add tolerations.
Change-Id: I8b3dacb71a7f102e7f74a6e4b6aee963ef12b8ed
This change allows us to substitute values into our rules files.
Example:
- alert: my_region_is_down
expr: up{region="{{ $my_region }}"} == 0
To support this change, rule annotations that used the expansion
{{ $labels.foo }} had to be surrounded with "{{` ... `}}" to render
correctly.
Change-Id: Ia7ac891de8261acca62105a3e2636bd747a5fbea
The patch moves dependency for job_rabbit_init from static to
dynamic section because of the fact that in some cases users and
credentials for OpenStack services had been created in RabbitMQ
before deployment of the OpepnStack chart for an service. Values is
going to have the below structure. cinder-rabbit-init could be moved
from static to dynamic section:
values:
dependencies:
dynamic:
job_rabbit_init:
api:
jobs:
- cinder-rabbit-init
Change-Id: Ib2b9858262a229390f775ad831f8c50dfb4a19da
The PS updates wait_for_pods function and adds query to filter the pods
which are not in Running or Succeeded state.
Also the PS reduces the amount of 'kubectl get' requests.
Change-Id: Ie2abdaf0a87ca377f5ce287a3de9e87d1ca6c0d4
Pass parameter from job allowing to parallelize helm tests using
separate scripts.
Change-Id: I3e06c5590d51c75448dc5ff5978dc7fc90daca6f
Signed-off-by: Andrii Ostapenko <andrii.ostapenko@att.com>
With this commit minikube is installed using contents of precreated
minikube-aio image containing installation script, all required binaries
and images inside. Pulling a single image from dockerhub via opendev
dockerhub proxy and loading images allows to save up to 6 minutes in
minikube installation.
Change-Id: I5936f440eb0567b8dcba2fdae614e4c5e88a7b9a
Signed-off-by: Andrii Ostapenko <andrii.ostapenko@att.com>
This chart could deploy fluentd either as a Deployment
or a Daemonset. Both options would use the deployment-fluentd
template with various sections toggled off based on values.yaml
I'd like to know - Does anyone run this chart as a Deployment?
We can simplify the chart, and zuul gates, by changing the chart
to deploy a Daemonset specifically.
Change-Id: Ie88ceadbf5113fc60e5bb0ddef09e18fe07a192c