| Commit message (Collapse) | Author | Age | Files | Lines |
|\
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
Automatic merge from submit-queue.
Install base_packages earlier
Currently, openshift_facts requires pyyaml to be installed.
This package is installed via init/base_packages.yml, which
is currently called after init/facts.yml. This results
in a situation where installs will fail due to missing
python dependency.
This commit splits init/facts.yml into two, and
allows base_packages.yml to be run before the
openshift_facts.py plugin is executed.
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
Currently, openshift_facts requires pyyaml to be installed.
This package is installed via init/base_packages.yml, which
is currently called after init/facts.yml. This results
in a situation where installs will fail due to missing
python dependency.
This commit splits init/facts.yml into two, and
allows base_packages.yml to be run before the
openshift_facts.py plugin is executed.
|
|/
|
|
|
|
|
| |
Allow playbooks/openshift-master/scaleup.yml to call
prerequisites.yml at the proper time.
Related-to: https://github.com/openshift/openshift-ansible/pull/6784
|
|\
| |
| | |
Fix node scaleup plays
|
| |
| |
| |
| |
| |
| |
| |
| | |
Currently, users have no way to run preqrequisites.yml
on just newly added nodes during scaleup.
This commit ensures only the new nodes are changed during
scaleup as well as ensure prerequisites are run.
|
|\ \
| | |
| | | |
Use master group's ansible_ssh_user when accepting nodes.
|
| | |
| | |
| | |
| | | |
that we can use the first master's ansible_ssh_user when delegating.
|
| | |
| | |
| | |
| | |
| | |
| | | |
The ansible-playbook command in the OpenStack readme is missing a
trailing backslash after the dynamic inventory. This prevents being able
to copy/paste the command into the terminal and just run it as is.
|
|\ \ \
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | | |
Automatic merge from submit-queue.
Drop the testing repo var from openstack readme
This was introduced as a temporary fix for the origin+centos
deployments. It is no longer necessary and it's confusing to
newcomers.
|
| | |/
| |/|
| | |
| | |
| | |
| | | |
This was introduced as a temporary fix for the origin+centos
deployments. It is no longer necessary and it's confusing to
newcomers.
|
|\ \ \
| |/ /
|/| |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | | |
Automatic merge from submit-queue.
Ensure atomic_proxies are configured with docker
In some setups, system containers may be used
in conjunction with package_docker.
This commit ensures that the atomic
proxies are configured while using package_docker
and containerized components.
This commit also ensures containerized_groups
are based on openshift_is_containerized instead
of 'containerized' variable directly to
account for possible use of atomic hosts which
users may not specify containerized in inventory
directly.
Fixes: https://bugzilla.redhat.com/show_bug.cgi?id=1535323
|
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | | |
In some setups, system containers may be used
in conjunction with package_docker.
This commit ensures that the atomic
proxies are configured while using package_docker
and containerized components.
This commit also ensures containerized_groups
are based on openshift_is_containerized instead
of 'containerized' variable directly to
account for possible use of atomic hosts which
users may not specify containerized in inventory
directly.
Fixes: https://bugzilla.redhat.com/show_bug.cgi?id=1535323
|
|\ \ \
| | | |
| | | | |
Bug 1511357: Task "Reload generated facts" failed while scaling up containerized masters
|
| |/ / |
|
|\ \ \
| | | |
| | | | |
Add new grafana playbook.
|
| | |/
| |/|
| | |
| | |
| | |
| | |
| | | |
this patch introduce a new playbook and roles to deploy grafana automaticly
using openshift-ansible capabilities.
this patch will abstract the grafana deployment process specially for openshift.
|
|\ \ \
| | | |
| | | | |
openshift_checks: repair adhoc list-checks mode
|
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | | |
fixes bug https://bugzilla.redhat.com/show_bug.cgi?id=1509157
docker_image_availability.__init__ was loading a variable that of course
was not defined for localhost, which is how this play runs. Fixed that.
Would be nice if there were a more generic way to solve that class of
problem but I can't think of anything for now that isn't worse.
Also noticed that this play gathers facts for localhost. That's not
necessary for what it's doing, so disabled that.
|
|\ \ \ \
| |_|_|/
|/| | | |
Add init/main.yml to etc-upgrade
|
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | | |
Currently, etcd upgrade playbook only calls
evaluate_groups to limit the number of tasks
that are run against nodes during initialization.
Recently, changes have been added to allow skipping
of most node tasks via providing variablized host
groups during initialization code.
This commit allows etcd upgrades to use common
init code to ensure necessary facts are defined
and proper variables are sanity checked.
Fixes: https://bugzilla.redhat.com/show_bug.cgi?id=1527771
|
|\ \ \ \
| |/ / /
|/| | | |
Faciliate 3.7 to 3.8 to 3.9 upgrade
|
| | | | |
|
| | | | |
|
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | | |
There are some obsoletes in 3.9 packages which lead yum to upgrade
directly from 3.7 to 3.9 when you ask for 3.8 packages while 3.9 repos
are enabled. Since we'd like to allow people to run one playbook to
upgrade from 3.7 to 3.8 to 3.9 we need to exclude those packages when
upgrading to 3.8
|
| | | | |
|
|\ \ \ \
| |/ / /
|/| | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | | |
Automatic merge from submit-queue.
Move OpenStack inventory.py out of sample-inventory
This moves the dynamic inventory for OpenStack from the `sample-inventory` dir users are expected to copy and edit to `/inventory/openstack`.
This will ensure that people don't use an outdated inventory as well as communicates that the inventory (similar to the playbooks and roles) is not something they are expected to modify under normal circumstances.
The sample inventory will now contain only `group_vars` and as such, moving from that to providing vars explicitly is also more obvious (you replace the `-i inventory` command with `-e @vars.yaml` and you're good to go).
|
| | | | |
|
| | |/
| |/|
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | | |
Keeping the dynamic inventory in the `sample-inventory` alongside the
`group_vars` poses a problem: when we update the inventory, our users
won't get the latest version.
The dynamic inventory should be independent of the OpenShift or
OpenStack configuration and the users should be able to either pass it
explicitly or create a symlink if they want.
|
|\ \ \
| |/ /
|/| | |
fix bug 1534271
|
| |/ |
|
|\ \
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | | |
Automatic merge from submit-queue.
Separate ELB & S3 from master node group provisioning
Moved s3 and elb provisioning out of `roles/openshift_aws/tasks/provisioning.yml` and into their own playbooks. These playbooks are now included in the provision playbook and in a WIP infrastructure playbook we intend to run up front in place of prerequisites.
@kwoodson what are your thoughts on something like this?
|
| | | |
|
| | |
| | |
| | |
| | | |
outside of the openshift_aws master provisioning tasks.
|
| | |
| | |
| | |
| | | |
Signed-off-by: Bogdan Dobrelya <bdobreli@redhat.com>
|
| |/
|/|
| |
| | |
Signed-off-by: Bogdan Dobrelya <bdobreli@redhat.com>
|
|\ \
| | |
| | | |
Install web console on upgrade
|
| | | |
|
|\ \ \
| | | |
| | | |
| | | |
| | | |
| | | |
| | | | |
Automatic merge from submit-queue.
Spelling and grammar changes to the advanced-configuration.md file.
I noticed some spelling errors when trying to read the OpenStack `advanced_configuration.md` file so I wanted to contribute the fixed spelling.
|
| | |/
| |/| |
|
|\ \ \
| |_|/
|/| | |
Move more plugins to lib_utils
|
| |/
| |
| |
| |
| |
| |
| | |
This commit continues moving plugins into lib_utils.
This commit does not move any plugins for add-on roles
such as logging and metrics.
|
|\ \
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | | |
Automatic merge from submit-queue.
Fix Cinder Persistent Volume support
This documents how to use Cinder-backed persistent volumes with OpenStack.
It needed a change to the dynamic inventory because the "openstack" cloudprovider plugin does actually require internal name resolution -- and the `openshift_hostname` value must match the name of the Nova server.
In addition, we need to be able to specify the V2 of the Cinder API for now as described in: https://github.com/openshift/openshift-docs/issues/5730
|
| | | |
|
| | | |
|
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | | |
When deploying on OpenStack with internal DNS configured, this will set
`openshift_hostname` to the Nova server name instead of its IP address.
Without those two matching, the OpenStack cloud provider configuration
will fail and the OpenShift nodes will not start.
|
| | |
| | |
| | |
| | |
| | |
| | |
| | | |
As described in[1], OpenShift currently only works with Block Storage
API v2 and the version autodetection is failing to figure that out.
[1]: https://github.com/openshift/openshift-docs/issues/5730
|
|\ \ \
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | | |
Automatic merge from submit-queue.
Add the ability to specify a timeout for node drain operations
A timeout to wait for nodes to drain pods can be specified to ensure that the upgrade continues even if nodes fail to drain pods in the allowed time. The default value of 0 will wait indefinitely allowing the admin to investigate the root cause and ensuring that disruption budgets are respected. In practice the `oc adm drain` command will eventually error out, at least that's what we've seen in our large online clusters, when that happens a second attempt will be made to drain the nodes, if it fails again it will abort the upgrade for that node or for the entire cluster based on your defined `openshift_upgrade_nodes_max_fail_percentage`.
`openshift_upgrade_nodes_drain_timeout=0` is the default and will wait until all pods have been drained successfully
`openshift_upgrade_nodes_drain_timeout=600` would wait for 600s before moving on to the tasks which would forcefully stop pods such as stopping docker, node, and openvswitch.
|
| | | | |
|
|\ \ \ \
| | | | |
| | | | | |
Ensure that openshift_facts role is imported whenever we rely on
|
| | | | |
| | | | |
| | | | |
| | | | | |
openshift_client_binary
|