| Commit message (Collapse) | Author | Age | Files | Lines |
|
|
|
|
|
| |
openshift_health_checker requires openshift_version, and that has been
updated to require group evaluation.
Update check playbooks and integration tests to do group eval.
|
| |
|
|\
| |
| | |
Restart NetworkManager only if dnsmasq was used
|
| | |
|
|\ \
| | |
| | | |
remove extra close brace in example inventory
|
| | | |
|
|\ \ \
| | | |
| | | | |
Remove package_update from install playbook
|
| | |/
| |/|
| | |
| | |
| | |
| | | |
We observed in our CI environment that this check can't guarantee that
an install would fail when the check fails, thus unless we can make its
output match the fate of an install we shall keep it disabled.
|
|/ / |
|
|/
|
|
|
|
| |
Created by command:
/usr/bin/tito tag --accept-auto-changelog --use-version=3.6.122
|
|
|
|
|
|
| |
Created by command:
/usr/bin/tito tag --accept-auto-changelog --use-version=3.6.121
|
|\
| |
| | |
Updating default from null to ""
|
|/ |
|
|
|
|
|
|
| |
Created by command:
/usr/bin/tito tag --accept-auto-changelog --use-version=3.6.120
|
|\
| |
| | |
Merged by openshift-bot
|
| | |
|
| |
| |
| |
| |
| | |
We cannot assume that 3.5 to 3.6 upgrades were signed with the correct
certs
|
| |
| |
| |
| |
| |
| |
| | |
We need to sort out how to know that the registry certificate has the
proper hostnames attached to it. It will for 3.6 clean installs but not
for 3.5 to 3.6 upgrades. For now make it opt in and come back to
this.
|
| | |
|
| |
| |
| |
| |
| |
| | |
Configures OPENSHIFT_DEFAULT_REGISTRY=docker-registry.default.svc
Adds 'cluster.local' to dns search on nodes via dispatcher script
Adds '.svc' to NO_PROXY defaults
|
| |
| |
| |
| |
| |
| | |
Created by command:
/usr/bin/tito tag --accept-auto-changelog --use-version=3.6.119
|
|\ \
| | |
| | | |
Merged by openshift-bot
|
| | |
| | |
| | |
| | |
| | |
| | |
| | | |
Some registries are not configured with valid certificates and thus the
check fails with 'http: server gave HTTP response to HTTPS client'.
Since this is not fetching images, but only checking for existence,
trade security for convenience.
|
|\ \ \
| | | |
| | | | |
Merged by openshift-bot
|
| | | |
| | | |
| | | |
| | | |
| | | | |
This would be the case if for instance they'd upgraded and then
migrated.
|
|\ \ \ \
| | | | |
| | | | | |
Merged by openshift-bot
|
| | | | |
| | | | |
| | | | |
| | | | | |
Signed-off-by: Jose A. Rivera <jarrpa@redhat.com>
|
| | | | |
| | | | |
| | | | |
| | | | | |
Signed-off-by: Jose A. Rivera <jarrpa@redhat.com>
|
| | | | |
| | | | |
| | | | |
| | | | | |
Signed-off-by: Jose A. Rivera <jarrpa@redhat.com>
|
| | | | |
| | | | |
| | | | |
| | | | | |
Signed-off-by: Jose A. Rivera <jarrpa@redhat.com>
|
| | | | |
| | | | |
| | | | |
| | | | | |
Signed-off-by: Jose A. Rivera <jarrpa@redhat.com>
|
| | | | |
| | | | |
| | | | |
| | | | | |
Signed-off-by: Jose A. Rivera <jarrpa@redhat.com>
|
| | | | |
| | | | |
| | | | |
| | | | | |
Signed-off-by: Jose A. Rivera <jarrpa@redhat.com>
|
|\ \ \ \ \
| | | | | |
| | | | | | |
Temporarilly only migrate jobs as we were before
|
| | |_|_|/
| |/| | | |
|
| | | | |
| | | | |
| | | | |
| | | | |
| | | | |
| | | | | |
Created by command:
/usr/bin/tito tag --accept-auto-changelog --use-version=3.6.117
|
|\ \ \ \ \
| |/ / / /
|/| | | | |
Merged by openshift-bot
|
| | | | | |
|
|\ \ \ \ \
| |_|_|/ /
|/| | | | |
Run storage upgrade pre and post master upgrade
|
| | | | | |
|
|\ \ \ \ \
| |_|_|_|/
|/| | | | |
Merged by openshift-bot
|
| | | | | |
|
| | | | |
| | | | |
| | | | |
| | | | |
| | | | |
| | | | | |
Created by command:
/usr/bin/tito tag --accept-auto-changelog --use-version=3.6.116
|
|\ \ \ \ \
| | | | | |
| | | | | | |
Merged by openshift-bot
|
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | | |
If we have no master config assume that we're a clean install.
If we're a clean install and we're 3.6 or greater use etcd v3 storage.
|
|\ \ \ \ \ \
| | | | | | |
| | | | | | | |
Merged by openshift-bot
|
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | | |
bug 1460564. Fixes [BZ #1460564](https://bugzilla.redhat.com/show_bug.cgi?id=1460564).
Unfortunately, the defaults for Elasticsearch prior to v5 allow more
than one "node" to access the same configured storage volume(s).
This change forces this value to 1 to ensure we don't have an ES pod
starting up accessing a volume while another ES pod is shutting down
when reploying. This can lead to "1" directories being created in
`/elasticsearch/persistent/${CLUSTER_NAME}/data/${CLUSTER_NAME}/nodes/`.
By default ES uses a "0" directory there when only one node is accessing
it.
|
|\ \ \ \ \ \ \
| |_|_|/ / / /
|/| | | | | | |
Rename cockpit-shell -> cockpit-system
|
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | | |
This is a temporary workaround to make sure the action plugin in the
openshift_health_checker role can read the correct value specified in
the OPENSHIFT_IMAGE_TAG environment variable, instead of a jinja2
template expression.
|
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | | |
The package name has changed.
See
https://bugzilla.redhat.com/show_bug.cgi?id=1461689
https://bugzilla.redhat.com/show_bug.cgi?id=1419718
|