| Commit message (Collapse) | Author | Age | Files | Lines |
| |
|
| |
|
|
|
|
|
|
| |
Fix embarrassing typo
fix etcd command
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
* add ansible-service-broker role to catalog install playbook
add ansible_service_broker to catalog install flow
fix indentation syntax error
respond to code review
Remove content, update image prefix to exclude image name
Make variables deployment type specific, add rhcc configuration, add fact validation
some spacing/comments
Remove ansible_service_broker_namespace variable
Fix stupid typo
== nil -> is none
nil -> null, fix validation checks
Fix a few typos
Add Broker object to catalog
replace oc_pvc creation with oc_obj
remove namespace from broker creation
fix namespace
|
| |
|
|\
| |
| | |
Merged by openshift-bot
|
| | |
|
| |
| |
| |
| |
| | |
We cannot assume that 3.5 to 3.6 upgrades were signed with the correct
certs
|
| |
| |
| |
| |
| |
| |
| | |
We need to sort out how to know that the registry certificate has the
proper hostnames attached to it. It will for 3.6 clean installs but not
for 3.5 to 3.6 upgrades. For now make it opt in and come back to
this.
|
| | |
|
| |
| |
| |
| |
| |
| | |
Configures OPENSHIFT_DEFAULT_REGISTRY=docker-registry.default.svc
Adds 'cluster.local' to dns search on nodes via dispatcher script
Adds '.svc' to NO_PROXY defaults
|
|\ \
| | |
| | | |
Merged by openshift-bot
|
| | |
| | |
| | |
| | |
| | |
| | |
| | | |
Some registries are not configured with valid certificates and thus the
check fails with 'http: server gave HTTP response to HTTPS client'.
Since this is not fetching images, but only checking for existence,
trade security for convenience.
|
|\ \ \
| | | |
| | | | |
Merged by openshift-bot
|
| | |/
| |/|
| | |
| | |
| | | |
This would be the case if for instance they'd upgraded and then
migrated.
|
|\ \ \
| | | |
| | | | |
Merged by openshift-bot
|
| | | |
| | | |
| | | |
| | | | |
Signed-off-by: Jose A. Rivera <jarrpa@redhat.com>
|
| | | |
| | | |
| | | |
| | | | |
Signed-off-by: Jose A. Rivera <jarrpa@redhat.com>
|
| | | |
| | | |
| | | |
| | | | |
Signed-off-by: Jose A. Rivera <jarrpa@redhat.com>
|
| | | |
| | | |
| | | |
| | | | |
Signed-off-by: Jose A. Rivera <jarrpa@redhat.com>
|
| | | |
| | | |
| | | |
| | | | |
Signed-off-by: Jose A. Rivera <jarrpa@redhat.com>
|
| | | |
| | | |
| | | |
| | | | |
Signed-off-by: Jose A. Rivera <jarrpa@redhat.com>
|
| | | |
| | | |
| | | |
| | | | |
Signed-off-by: Jose A. Rivera <jarrpa@redhat.com>
|
|\ \ \ \
| |_|/ /
|/| | | |
Merged by openshift-bot
|
| | | | |
|
|\ \ \ \
| |_|_|/
|/| | | |
Merged by openshift-bot
|
| | | | |
|
|\ \ \ \
| | | | |
| | | | | |
Merged by openshift-bot
|
| | | | |
| | | | |
| | | | |
| | | | |
| | | | | |
If we have no master config assume that we're a clean install.
If we're a clean install and we're 3.6 or greater use etcd v3 storage.
|
|\ \ \ \ \
| | | | | |
| | | | | | |
Merged by openshift-bot
|
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | | |
bug 1460564. Fixes [BZ #1460564](https://bugzilla.redhat.com/show_bug.cgi?id=1460564).
Unfortunately, the defaults for Elasticsearch prior to v5 allow more
than one "node" to access the same configured storage volume(s).
This change forces this value to 1 to ensure we don't have an ES pod
starting up accessing a volume while another ES pod is shutting down
when reploying. This can lead to "1" directories being created in
`/elasticsearch/persistent/${CLUSTER_NAME}/data/${CLUSTER_NAME}/nodes/`.
By default ES uses a "0" directory there when only one node is accessing
it.
|
|\ \ \ \ \ \
| |_|_|/ / /
|/| | | | | |
Rename cockpit-shell -> cockpit-system
|
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | | |
The package name has changed.
See
https://bugzilla.redhat.com/show_bug.cgi?id=1461689
https://bugzilla.redhat.com/show_bug.cgi?id=1419718
|
|\ \ \ \ \ \
| |/ / / / /
|/| | | | | |
Update CloudForms templates for CF 4.5/CF 4.2
|
| | | | | |
| | | | | |
| | | | | |
| | | | | | |
'cloudforms42' for CF 4.2.
|
| | | | | |
| | | | | |
| | | | | |
| | | | | | |
'cloudforms45' for CF 4.5.
|
| | | | | | |
|
|\ \ \ \ \ \
| | | | | | |
| | | | | | | |
Merged by openshift-bot
|
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | | |
port the code that creates the external Elasticsearch routes to the
new logging roles
Have to suppress this error message:
SSL Problem illegal change cipher spec msg, conn state = 6, handshake state = 1
which is coming from the router health check, until
https://github.com/openshift/origin/issues/14515
is fixed - otherwise, the es log is spammed relentlessly
|
|\ \ \ \ \ \ \
| | | | | | | |
| | | | | | | | |
Merged by openshift-bot
|
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | | |
We cannot rely on the `watch.Until` call in the `rollout status`
subcommand for the time being, so we need to ignore the result of this
call. This will make the rollout status check best-effort, so we need to
follow it with a poll for the actual status of the rollout, which we can
extract from the `openshift.io/deployment.phase` annotation on the
ReplicationControllers. This annotation can have only three values --
`Running`, `Complete` and `Failed`. If we poll on this attribute until
we stop seeing `Running`, we can then inspect the last result for
`Failed`; if it's present, we have failed the deployment.
Signed-off-by: Steve Kuznetsov <skuznets@redhat.com>
|
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | | |
When deploying the hosted router and registry components, we need to
ensure that they correctly roll out. The previous checks were weak in
that they either simply waited for a set amount of time and/or did one
replica check. They would fail if the router or registry took longer to
deploy or if there were un-ready or failing replicas. The `oc rollout`
command group contains the `status` endpoint for internalizing all of
the logic for determining when a rollout has succeeded or failed, so
simply using this client call will ensure that the router and registry
correctly deploy.
Signed-off-by: Steve Kuznetsov <skuznets@redhat.com>
|
|\ \ \ \ \ \ \ \
| |_|_|_|/ / / /
|/| | | | | | | |
CFME as an OpenShift Pod
|
| | | | | | | | |
|
| | | | | | | | |
|
| | | | | | | | |
|
| | |_|_|_|_|/
| |/| | | | | |
|
|\ \ \ \ \ \ \
| | | | | | | |
| | | | | | | | |
Merged by openshift-bot
|
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | | |
adding
openshift_logging_fluentd_buffer_queue_limit: 1024
openshift_logging_fluentd_buffer_size_limit: 1m
openshift_logging_mux_buffer_queue_limit: 1024
openshift_logging_mux_buffer_size_limit: 1m
and setting the matched environment variables.
|
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | | |
roles/openshift_logging_fluentd/defaults/main.yml.
Adding the description for openshift_hosted_logging_fluentd_{cpu,memory}_limit to README.md.
Replaced openshift_hosted_logging_mux_.* with openshift_logging_mux_.*.
|