ansible-playbook 2.9.27 config file = None configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/local/lib/python3.9/site-packages/ansible executable location = /usr/local/bin/ansible-playbook python version = 3.9.19 (main, May 16 2024, 11:40:09) [GCC 8.5.0 20210514 (Red Hat 8.5.0-22)] No config file found; using defaults [WARNING]: running playbook inside collection fedora.linux_system_roles statically imported: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml statically imported: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_test_data.yml statically imported: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/check_candlepin.yml Skipping callback 'actionable', as we already have a stdout callback. Skipping callback 'counter_enabled', as we already have a stdout callback. Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'dense', as we already have a stdout callback. Skipping callback 'dense', as we already have a stdout callback. Skipping callback 'full_skip', as we already have a stdout callback. Skipping callback 'json', as we already have a stdout callback. Skipping callback 'jsonl', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'null', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. Skipping callback 'selective', as we already have a stdout callback. Skipping callback 'skippy', as we already have a stdout callback. Skipping callback 'stderr', as we already have a stdout callback. Skipping callback 'unixy', as we already have a stdout callback. Skipping callback 'yaml', as we already have a stdout callback. PLAYBOOK: tests_environments.yml *********************************************** 1 plays in /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tests_environments.yml PLAY [Basic repository enablement/disablement test] **************************** META: ran handlers TASK [Get LSR_RHC_TEST_DATA environment variable] ****************************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_test_data.yml:3 Saturday 08 November 2025 14:20:51 -0500 (0:00:00.021) 0:00:00.021 ***** ok: [managed-node2] => { "ansible_facts": { "lsr_rhc_test_data_file": "" }, "changed": false } TASK [Import test data] ******************************************************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_test_data.yml:12 Saturday 08 November 2025 14:20:51 -0500 (0:00:00.043) 0:00:00.065 ***** skipping: [managed-node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get facts for external test data] **************************************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_test_data.yml:16 Saturday 08 November 2025 14:20:51 -0500 (0:00:00.033) 0:00:00.098 ***** skipping: [managed-node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set local lsr_rhc_test_data] ********************************************* task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_test_data.yml:24 Saturday 08 November 2025 14:20:51 -0500 (0:00:00.030) 0:00:00.128 ***** ok: [managed-node2] => { "ansible_facts": { "lsr_rhc_test_data": { "baseurl": "http://localhost:8080", "candlepin_host": "candlepin.local", "candlepin_insecure": false, "candlepin_port": 8443, "candlepin_prefix": "/candlepin", "env_nonworking": "Ceci n'est pas une environment", "envs_register": [ "Environment 2" ], "insights": false, "proxy_auth_hostname": "localhost", "proxy_auth_password": "proxypass", "proxy_auth_port": 3130, "proxy_auth_scheme": "https", "proxy_auth_username": "proxyuser", "proxy_noauth_hostname": "localhost", "proxy_noauth_port": 3128, "proxy_noauth_scheme": "https", "proxy_nonworking_hostname": "wrongproxy", "proxy_nonworking_password": "wrong-proxypassword", "proxy_nonworking_port": 4000, "proxy_nonworking_username": "wrong-proxyuser", "reg_activation_keys": [ "default_key" ], "reg_invalid_password": "invalid-password", "reg_invalid_username": "invalid-user", "reg_organization": "donaldduck", "reg_password": "password", "reg_username": "doc", "release": null, "repositories": [ { "name": "donaldy-content-label-7051", "state": "enabled" }, { "name": "content-label-32060", "state": "disabled" } ] } }, "ansible_included_var_files": [ "/tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/../files/candlepin_data.yml" ], "changed": false } TASK [Check if system is ostree] *********************************************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_test_data.yml:32 Saturday 08 November 2025 14:20:51 -0500 (0:00:00.033) 0:00:00.162 ***** ok: [managed-node2] => { "ansible_facts": { "discovered_interpreter_python": "/usr/libexec/platform-python" }, "changed": false, "stat": { "exists": false } } TASK [Set flag to indicate system is ostree] *********************************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_test_data.yml:37 Saturday 08 November 2025 14:20:51 -0500 (0:00:00.664) 0:00:00.826 ***** ok: [managed-node2] => { "ansible_facts": { "__rhc_is_ostree": false }, "changed": false } TASK [Get facts for external test data] **************************************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:9 Saturday 08 November 2025 14:20:51 -0500 (0:00:00.034) 0:00:00.861 ***** ok: [managed-node2] TASK [Set helper fact for Candlepin base URL] ********************************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:17 Saturday 08 November 2025 14:20:52 -0500 (0:00:00.619) 0:00:01.480 ***** ok: [managed-node2] => { "ansible_facts": { "_cp_url": "https://candlepin.local:8443/candlepin" }, "changed": false } TASK [Set helper fact for Candlepin owner URL] ********************************* task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:21 Saturday 08 November 2025 14:20:52 -0500 (0:00:00.035) 0:00:01.516 ***** ok: [managed-node2] => { "ansible_facts": { "_cp_url_owner": "https://candlepin.local:8443/candlepin/owners/donaldduck" }, "changed": false } TASK [Add candlepin hostname to /etc/hosts] ************************************ task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:25 Saturday 08 November 2025 14:20:52 -0500 (0:00:00.033) 0:00:01.549 ***** changed: [managed-node2] => { "backup": "", "changed": true } MSG: line added TASK [Install needed packages] ************************************************* task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:31 Saturday 08 November 2025 14:20:53 -0500 (0:00:00.444) 0:00:01.994 ***** changed: [managed-node2] => { "changed": true, "rc": 0, "results": [ "Installed: shadow-utils-subid-2:4.6-22.el8.x86_64", "Installed: container-selinux-2:2.229.0-2.module_el8+847+7863d4e6.noarch", "Installed: criu-3.18-4.module_el8+804+f131391c.x86_64", "Installed: dnsmasq-2.79-33.el8.x86_64", "Installed: libslirp-4.4.0-1.module_el8+804+f131391c.x86_64", "Installed: protobuf-c-1.3.0-8.el8.x86_64", "Installed: slirp4netns-1.2.3-1.module_el8+951+32019cde.x86_64", "Installed: podman-catatonit-3:4.9.4-0.1.module_el8+971+3d3df00d.x86_64", "Installed: fuse3-libs-3.3.0-19.el8.x86_64", "Installed: fuse3-3.3.0-19.el8.x86_64", "Installed: containers-common-2:1-81.module_el8+968+fbb249c7.x86_64", "Installed: podman-gvproxy-3:4.9.4-0.1.module_el8+971+3d3df00d.x86_64", "Installed: podman-plugins-3:4.9.4-0.1.module_el8+971+3d3df00d.x86_64", "Installed: fuse-overlayfs-1.13-1.module_el8+804+f131391c.x86_64", "Installed: fuse-common-3.3.0-19.el8.x86_64", "Installed: podman-3:4.9.4-0.1.module_el8+971+3d3df00d.x86_64", "Installed: libnet-1.1.6-15.el8.x86_64", "Installed: runc-1:1.1.12-1.module_el8+885+7da147f3.x86_64", "Installed: containernetworking-plugins-1:1.4.0-2.module_el8+974+0c52b299.x86_64", "Installed: conmon-3:2.1.10-1.module_el8+804+f131391c.x86_64" ] } lsrpackages: podman TASK [Clean up Candlepin container] ******************************************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:38 Saturday 08 November 2025 14:21:53 -0500 (0:01:00.720) 0:01:02.714 ***** included: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/teardown_candlepin.yml for managed-node2 TASK [Check if the candlepin container exists] ********************************* task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/teardown_candlepin.yml:6 Saturday 08 November 2025 14:21:53 -0500 (0:00:00.040) 0:01:02.754 ***** ok: [managed-node2] => { "changed": false, "cmd": [ "podman", "ps", "-a", "--filter", "name=candlepin" ], "delta": "0:00:01.570791", "end": "2025-11-08 14:21:55.869592", "rc": 0, "start": "2025-11-08 14:21:54.298801" } STDOUT: CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES TASK [Ensure that Candlepin container doesn't exist] *************************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/teardown_candlepin.yml:17 Saturday 08 November 2025 14:21:55 -0500 (0:00:02.027) 0:01:04.781 ***** skipping: [managed-node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Start Candlepin container] *********************************************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:41 Saturday 08 November 2025 14:21:55 -0500 (0:00:00.034) 0:01:04.816 ***** ok: [managed-node2] => { "changed": false, "cmd": [ "podman", "run", "--rm", "--detach", "--hostname", "candlepin.local", "--name", "candlepin", "--publish", "8443:8443", "--publish", "8080:8080", "ghcr.io/candlepin/candlepin-unofficial" ], "delta": "0:00:15.546426", "end": "2025-11-08 14:22:11.803647", "rc": 0, "start": "2025-11-08 14:21:56.257221" } STDOUT: b9ded374cfc89cf2a47688a5a1eee040580bdb4398a1ca54f42d1325a4e44440 STDERR: Trying to pull ghcr.io/candlepin/candlepin-unofficial:latest... Getting image source signatures Copying blob sha256:868e32b0b96932a9d44af4fddb5291921afffb37e16e8c9bc0382ef20f02e4a0 Copying config sha256:fc49ff13f7f3d9b39189a4dadc708bc5cf2aea44997b748d698128d169c494b8 Writing manifest to image destination TASK [Ensure directories exist] ************************************************ task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:62 Saturday 08 November 2025 14:22:11 -0500 (0:00:15.932) 0:01:20.748 ***** changed: [managed-node2] => (item=/etc/pki/product) => { "ansible_loop_var": "item", "changed": true, "gid": 0, "group": "root", "item": "/etc/pki/product", "mode": "0755", "owner": "root", "path": "/etc/pki/product", "secontext": "unconfined_u:object_r:cert_t:s0", "size": 6, "state": "directory", "uid": 0 } changed: [managed-node2] => (item=/etc/pki/product-default) => { "ansible_loop_var": "item", "changed": true, "gid": 0, "group": "root", "item": "/etc/pki/product-default", "mode": "0755", "owner": "root", "path": "/etc/pki/product-default", "secontext": "unconfined_u:object_r:cert_t:s0", "size": 6, "state": "directory", "uid": 0 } changed: [managed-node2] => (item=/etc/rhsm/ca) => { "ansible_loop_var": "item", "changed": true, "gid": 0, "group": "root", "item": "/etc/rhsm/ca", "mode": "0755", "owner": "root", "path": "/etc/rhsm/ca", "secontext": "unconfined_u:object_r:rhsmcertd_config_t:s0", "size": 6, "state": "directory", "uid": 0 } TASK [Copy product certificates] *********************************************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:72 Saturday 08 November 2025 14:22:15 -0500 (0:00:03.772) 0:01:24.521 ***** ok: [managed-node2] => (item=7050) => { "ansible_loop_var": "item", "changed": false, "cmd": [ "podman", "cp", "candlepin:/home/candlepin/devel/candlepin/generated_certs/7050.pem", "/etc/pki/product-default/" ], "delta": "0:00:00.605901", "end": "2025-11-08 14:22:18.063147", "item": "7050", "rc": 0, "start": "2025-11-08 14:22:17.457246" } TASK [Copy Candlepin CA certificate for subscription-manager] ****************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:83 Saturday 08 November 2025 14:22:18 -0500 (0:00:02.550) 0:01:27.072 ***** ok: [managed-node2] => { "changed": false, "cmd": [ "podman", "cp", "candlepin:/etc/candlepin/certs/candlepin-ca.crt", "/etc/rhsm/ca/candlepin-ca.pem" ], "delta": "0:00:00.401232", "end": "2025-11-08 14:22:19.520044", "rc": 0, "start": "2025-11-08 14:22:19.118812" } TASK [Copy Candlepin CA certificate for system] ******************************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:92 Saturday 08 November 2025 14:22:19 -0500 (0:00:01.432) 0:01:28.504 ***** ok: [managed-node2] => { "changed": false, "cmd": [ "podman", "cp", "candlepin:/etc/candlepin/certs/candlepin-ca.crt", "/etc/pki/ca-trust/source/anchors/candlepin-ca.pem" ], "delta": "0:00:00.466693", "end": "2025-11-08 14:22:20.856268", "rc": 0, "start": "2025-11-08 14:22:20.389575" } TASK [Update system certificates store] **************************************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:101 Saturday 08 November 2025 14:22:20 -0500 (0:00:01.352) 0:01:29.857 ***** ok: [managed-node2] => { "changed": false, "cmd": [ "update-ca-trust", "extract" ], "delta": "0:00:01.385276", "end": "2025-11-08 14:22:23.010289", "rc": 0, "start": "2025-11-08 14:22:21.625013" } TASK [Wait for started Candlepin] ********************************************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:108 Saturday 08 November 2025 14:22:23 -0500 (0:00:02.119) 0:01:31.977 ***** ok: [managed-node2] => { "attempts": 1, "changed": false, "connection": "close", "content_type": "application/json", "cookies": {}, "cookies_string": "", "date": "Sat, 08 Nov 2025 19:22:37 GMT", "elapsed": 12, "redirected": true, "status": 200, "transfer_encoding": "chunked", "url": "https://candlepin.local:8443/candlepin/", "vary": "accept-encoding", "x_candlepin_request_uuid": "0bb61fc8-c07a-429d-992e-70ce4172b121", "x_version": "4.7.1-1" } MSG: OK (unknown bytes) TASK [Install GPG key for RPM repositories] ************************************ task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:118 Saturday 08 November 2025 14:22:37 -0500 (0:00:14.565) 0:01:46.543 ***** changed: [managed-node2] => { "changed": true, "checksum_dest": null, "checksum_src": "5bd09883847285c54e6064f29dd9686c1afa5d72", "dest": "/etc/pki/rpm-gpg/RPM-GPG-KEY-candlepin", "elapsed": 0, "gid": 0, "group": "root", "md5sum": "aadf73f83655a28e287fab4099f1e17a", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:cert_t:s0", "size": 1660, "src": "/root/.ansible/tmp/ansible-tmp-1762629757.7391105-9999-41635758179647/tmp6itp5mz_", "state": "file", "status_code": 200, "uid": 0, "url": "http://candlepin.local:8080/RPM-GPG-KEY-candlepin" } MSG: OK (1660 bytes) TASK [Add environments] ******************************************************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:127 Saturday 08 November 2025 14:22:38 -0500 (0:00:00.535) 0:01:47.078 ***** ok: [managed-node2] => (item={'name': 'Environment 1', 'desc': 'The environment 1', 'id': 'envId1'}) => { "ansible_loop_var": "item", "changed": false, "connection": "close", "content_type": "application/json", "cookies": {}, "cookies_string": "", "date": "Sat, 08 Nov 2025 19:22:38 GMT", "elapsed": 0, "item": { "desc": "The environment 1", "id": "envId1", "name": "Environment 1" }, "json": { "contentPrefix": null, "created": "2025-11-08T19:22:39+0000", "description": "The environment 1", "environmentContent": [], "id": "envId1", "name": "Environment 1", "owner": { "anonymous": false, "contentAccessMode": "org_environment", "displayName": "Donald Duck", "href": "/owners/donaldduck", "id": "8a8082e4997b71a401997b71b9520004", "key": "donaldduck" }, "type": null, "updated": "2025-11-08T19:22:39+0000" }, "redirected": false, "status": 200, "transfer_encoding": "chunked", "url": "https://candlepin.local:8443/candlepin/owners/donaldduck/environments", "vary": "accept-encoding", "x_candlepin_request_uuid": "da59c6db-d7a8-4ef4-98e2-b0ef1d7f6f04", "x_version": "4.7.1-1" } MSG: OK (unknown bytes) ok: [managed-node2] => (item={'name': 'Environment 2', 'desc': 'The environment 2', 'id': 'envId2'}) => { "ansible_loop_var": "item", "changed": false, "connection": "close", "content_type": "application/json", "cookies": {}, "cookies_string": "", "date": "Sat, 08 Nov 2025 19:22:39 GMT", "elapsed": 0, "item": { "desc": "The environment 2", "id": "envId2", "name": "Environment 2" }, "json": { "contentPrefix": null, "created": "2025-11-08T19:22:39+0000", "description": "The environment 2", "environmentContent": [], "id": "envId2", "name": "Environment 2", "owner": { "anonymous": false, "contentAccessMode": "org_environment", "displayName": "Donald Duck", "href": "/owners/donaldduck", "id": "8a8082e4997b71a401997b71b9520004", "key": "donaldduck" }, "type": null, "updated": "2025-11-08T19:22:39+0000" }, "redirected": false, "status": 200, "transfer_encoding": "chunked", "url": "https://candlepin.local:8443/candlepin/owners/donaldduck/environments", "vary": "accept-encoding", "x_candlepin_request_uuid": "782a99c7-b062-4844-982b-179f931580bc", "x_version": "4.7.1-1" } MSG: OK (unknown bytes) TASK [Check Candlepin works] *************************************************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/check_candlepin.yml:3 Saturday 08 November 2025 14:22:39 -0500 (0:00:01.655) 0:01:48.734 ***** ok: [managed-node2] => { "changed": false, "connection": "close", "content_type": "application/json", "cookies": {}, "cookies_string": "", "date": "Sat, 08 Nov 2025 19:22:39 GMT", "elapsed": 0, "redirected": true, "status": 200, "transfer_encoding": "chunked", "url": "https://candlepin.local:8443/candlepin/", "vary": "accept-encoding", "x_candlepin_request_uuid": "13168c9a-f0b3-4d45-b046-e3f7bd2e573e", "x_version": "4.7.1-1" } MSG: OK (unknown bytes) META: TASK [Ensure ansible_facts used by the test] *********************************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tests_environments.yml:24 Saturday 08 November 2025 14:22:40 -0500 (0:00:00.508) 0:01:49.242 ***** ok: [managed-node2] TASK [Register (wrong environment)] ******************************************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tests_environments.yml:35 Saturday 08 November 2025 14:22:40 -0500 (0:00:00.424) 0:01:49.666 ***** TASK [fedora.linux_system_roles.rhc : Set ansible_facts required by role] ****** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/main.yml:3 Saturday 08 November 2025 14:22:40 -0500 (0:00:00.020) 0:01:49.686 ***** included: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/set_vars.yml for managed-node2 TASK [fedora.linux_system_roles.rhc : Ensure ansible_facts used by role] ******* task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/set_vars.yml:3 Saturday 08 November 2025 14:22:40 -0500 (0:00:00.046) 0:01:49.733 ***** skipping: [managed-node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.rhc : Check if system is ostree] *************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/set_vars.yml:11 Saturday 08 November 2025 14:22:40 -0500 (0:00:00.013) 0:01:49.747 ***** skipping: [managed-node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.rhc : Set flag to indicate system is ostree] *** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/set_vars.yml:16 Saturday 08 November 2025 14:22:40 -0500 (0:00:00.012) 0:01:49.760 ***** skipping: [managed-node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.rhc : Check if insights-packages are installed] *** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/set_vars.yml:20 Saturday 08 November 2025 14:22:40 -0500 (0:00:00.025) 0:01:49.786 ***** skipping: [managed-node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.rhc : Handle insights unregistration] ********** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/main.yml:6 Saturday 08 November 2025 14:22:40 -0500 (0:00:00.019) 0:01:49.805 ***** skipping: [managed-node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.rhc : Handle system subscription] ************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/main.yml:15 Saturday 08 November 2025 14:22:40 -0500 (0:00:00.017) 0:01:49.822 ***** included: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/subscription-manager.yml for managed-node2 TASK [fedora.linux_system_roles.rhc : Ensure required packages are installed] *** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/subscription-manager.yml:3 Saturday 08 November 2025 14:22:40 -0500 (0:00:00.022) 0:01:49.845 ***** changed: [managed-node2] => { "changed": true, "rc": 0, "results": [ "Installed: dnf-plugin-subscription-manager-1.28.42-1.el8.x86_64", "Installed: subscription-manager-rhsm-certificates-20220623-1.el8.noarch", "Installed: subscription-manager-1.28.42-1.el8.x86_64", "Installed: python3-iniparse-0.4-31.el8.noarch", "Installed: python3-inotify-0.9.6-13.el8.noarch", "Installed: python3-subscription-manager-rhsm-1.28.42-1.el8.x86_64", "Installed: python3-librepo-1.14.2-5.el8.x86_64", "Installed: python3-cloud-what-1.28.42-1.el8.x86_64", "Installed: usermode-1.113-2.el8.x86_64", "Installed: python3-ethtool-0.14-5.el8.x86_64" ] } lsrpackages: subscription-manager TASK [fedora.linux_system_roles.rhc : Get subscription status] ***************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/subscription-manager.yml:10 Saturday 08 November 2025 14:22:45 -0500 (0:00:04.510) 0:01:54.356 ***** skipping: [managed-node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.rhc : Call subscription-manager] *************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/subscription-manager.yml:23 Saturday 08 November 2025 14:22:45 -0500 (0:00:00.015) 0:01:54.371 ***** An exception occurred during task execution. To see the full traceback, use -vvv. The error was: SyntaxError: future feature annotations is not defined fatal: [managed-node2]: FAILED! => { "changed": false, "rc": 1 } MSG: MODULE FAILURE See stdout/stderr for the exact error MODULE_STDOUT: Traceback (most recent call last): File "/root/.ansible/tmp/ansible-tmp-1762629765.5851321-10161-74016621386902/AnsiballZ_redhat_subscription.py", line 102, in _ansiballz_main() File "/root/.ansible/tmp/ansible-tmp-1762629765.5851321-10161-74016621386902/AnsiballZ_redhat_subscription.py", line 94, in _ansiballz_main invoke_module(zipped_mod, temp_path, ANSIBALLZ_PARAMS) File "/root/.ansible/tmp/ansible-tmp-1762629765.5851321-10161-74016621386902/AnsiballZ_redhat_subscription.py", line 40, in invoke_module runpy.run_module(mod_name='ansible_collections.community.general.plugins.modules.redhat_subscription', init_globals=None, run_name='__main__', alter_sys=True) File "/usr/lib64/python3.6/runpy.py", line 201, in run_module mod_name, mod_spec, code = _get_module_details(mod_name) File "/usr/lib64/python3.6/runpy.py", line 128, in _get_module_details spec = importlib.util.find_spec(mod_name) File "/usr/lib64/python3.6/importlib/util.py", line 89, in find_spec return _find_spec(fullname, parent.__path__) File "", line 894, in _find_spec File "", line 1157, in find_spec File "", line 1131, in _get_spec File "", line 1112, in _legacy_get_spec File "", line 441, in spec_from_loader File "", line 544, in spec_from_file_location File "/tmp/ansible_community.general.redhat_subscription_payload_5olhs_pz/ansible_community.general.redhat_subscription_payload.zip/ansible_collections/community/general/plugins/modules/redhat_subscription.py", line 8 SyntaxError: future feature annotations is not defined MODULE_STDERR: Shared connection to 10.31.40.154 closed. TASK [Assert registration failed] ********************************************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tests_environments.yml:59 Saturday 08 November 2025 14:22:45 -0500 (0:00:00.475) 0:01:54.846 ***** ok: [managed-node2] => { "changed": false } MSG: All assertions passed TASK [Register (with existing environments)] *********************************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tests_environments.yml:65 Saturday 08 November 2025 14:22:45 -0500 (0:00:00.016) 0:01:54.863 ***** TASK [fedora.linux_system_roles.rhc : Set ansible_facts required by role] ****** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/main.yml:3 Saturday 08 November 2025 14:22:46 -0500 (0:00:00.017) 0:01:54.881 ***** included: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/set_vars.yml for managed-node2 TASK [fedora.linux_system_roles.rhc : Ensure ansible_facts used by role] ******* task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/set_vars.yml:3 Saturday 08 November 2025 14:22:46 -0500 (0:00:00.017) 0:01:54.898 ***** skipping: [managed-node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.rhc : Check if system is ostree] *************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/set_vars.yml:11 Saturday 08 November 2025 14:22:46 -0500 (0:00:00.013) 0:01:54.911 ***** skipping: [managed-node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.rhc : Set flag to indicate system is ostree] *** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/set_vars.yml:16 Saturday 08 November 2025 14:22:46 -0500 (0:00:00.012) 0:01:54.923 ***** skipping: [managed-node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.rhc : Check if insights-packages are installed] *** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/set_vars.yml:20 Saturday 08 November 2025 14:22:46 -0500 (0:00:00.012) 0:01:54.936 ***** skipping: [managed-node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.rhc : Handle insights unregistration] ********** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/main.yml:6 Saturday 08 November 2025 14:22:46 -0500 (0:00:00.011) 0:01:54.947 ***** skipping: [managed-node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.rhc : Handle system subscription] ************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/main.yml:15 Saturday 08 November 2025 14:22:46 -0500 (0:00:00.012) 0:01:54.960 ***** included: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/subscription-manager.yml for managed-node2 TASK [fedora.linux_system_roles.rhc : Ensure required packages are installed] *** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/subscription-manager.yml:3 Saturday 08 November 2025 14:22:46 -0500 (0:00:00.019) 0:01:54.980 ***** ok: [managed-node2] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do lsrpackages: subscription-manager TASK [fedora.linux_system_roles.rhc : Get subscription status] ***************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/subscription-manager.yml:10 Saturday 08 November 2025 14:22:48 -0500 (0:00:02.816) 0:01:57.796 ***** skipping: [managed-node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.rhc : Call subscription-manager] *************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/subscription-manager.yml:23 Saturday 08 November 2025 14:22:48 -0500 (0:00:00.018) 0:01:57.814 ***** An exception occurred during task execution. To see the full traceback, use -vvv. The error was: SyntaxError: future feature annotations is not defined fatal: [managed-node2]: FAILED! => { "changed": false, "rc": 1 } MSG: MODULE FAILURE See stdout/stderr for the exact error MODULE_STDOUT: Traceback (most recent call last): File "/root/.ansible/tmp/ansible-tmp-1762629769.027101-10242-824133678158/AnsiballZ_redhat_subscription.py", line 102, in _ansiballz_main() File "/root/.ansible/tmp/ansible-tmp-1762629769.027101-10242-824133678158/AnsiballZ_redhat_subscription.py", line 94, in _ansiballz_main invoke_module(zipped_mod, temp_path, ANSIBALLZ_PARAMS) File "/root/.ansible/tmp/ansible-tmp-1762629769.027101-10242-824133678158/AnsiballZ_redhat_subscription.py", line 40, in invoke_module runpy.run_module(mod_name='ansible_collections.community.general.plugins.modules.redhat_subscription', init_globals=None, run_name='__main__', alter_sys=True) File "/usr/lib64/python3.6/runpy.py", line 201, in run_module mod_name, mod_spec, code = _get_module_details(mod_name) File "/usr/lib64/python3.6/runpy.py", line 128, in _get_module_details spec = importlib.util.find_spec(mod_name) File "/usr/lib64/python3.6/importlib/util.py", line 89, in find_spec return _find_spec(fullname, parent.__path__) File "", line 894, in _find_spec File "", line 1157, in find_spec File "", line 1131, in _get_spec File "", line 1112, in _legacy_get_spec File "", line 441, in spec_from_loader File "", line 544, in spec_from_file_location File "/tmp/ansible_community.general.redhat_subscription_payload_uf_jjnn9/ansible_community.general.redhat_subscription_payload.zip/ansible_collections/community/general/plugins/modules/redhat_subscription.py", line 8 SyntaxError: future feature annotations is not defined MODULE_STDERR: Shared connection to 10.31.40.154 closed. TASK [Unregister] ************************************************************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tests_environments.yml:105 Saturday 08 November 2025 14:22:49 -0500 (0:00:00.364) 0:01:58.179 ***** TASK [fedora.linux_system_roles.rhc : Set ansible_facts required by role] ****** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/main.yml:3 Saturday 08 November 2025 14:22:49 -0500 (0:00:00.020) 0:01:58.199 ***** included: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/set_vars.yml for managed-node2 TASK [fedora.linux_system_roles.rhc : Ensure ansible_facts used by role] ******* task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/set_vars.yml:3 Saturday 08 November 2025 14:22:49 -0500 (0:00:00.016) 0:01:58.215 ***** skipping: [managed-node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.rhc : Check if system is ostree] *************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/set_vars.yml:11 Saturday 08 November 2025 14:22:49 -0500 (0:00:00.013) 0:01:58.229 ***** skipping: [managed-node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.rhc : Set flag to indicate system is ostree] *** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/set_vars.yml:16 Saturday 08 November 2025 14:22:49 -0500 (0:00:00.012) 0:01:58.242 ***** skipping: [managed-node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.rhc : Check if insights-packages are installed] *** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/set_vars.yml:20 Saturday 08 November 2025 14:22:49 -0500 (0:00:00.011) 0:01:58.254 ***** skipping: [managed-node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.rhc : Handle insights unregistration] ********** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/main.yml:6 Saturday 08 November 2025 14:22:49 -0500 (0:00:00.012) 0:01:58.266 ***** skipping: [managed-node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.rhc : Handle system subscription] ************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/main.yml:15 Saturday 08 November 2025 14:22:49 -0500 (0:00:00.011) 0:01:58.278 ***** included: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/subscription-manager.yml for managed-node2 TASK [fedora.linux_system_roles.rhc : Ensure required packages are installed] *** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/subscription-manager.yml:3 Saturday 08 November 2025 14:22:49 -0500 (0:00:00.019) 0:01:58.298 ***** skipping: [managed-node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.rhc : Get subscription status] ***************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/subscription-manager.yml:10 Saturday 08 November 2025 14:22:49 -0500 (0:00:00.011) 0:01:58.310 ***** skipping: [managed-node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.rhc : Call subscription-manager] *************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/subscription-manager.yml:23 Saturday 08 November 2025 14:22:49 -0500 (0:00:00.011) 0:01:58.322 ***** An exception occurred during task execution. To see the full traceback, use -vvv. The error was: SyntaxError: future feature annotations is not defined fatal: [managed-node2]: FAILED! => { "changed": false, "rc": 1 } MSG: MODULE FAILURE See stdout/stderr for the exact error MODULE_STDOUT: Traceback (most recent call last): File "/root/.ansible/tmp/ansible-tmp-1762629769.5202293-10278-20579478037084/AnsiballZ_redhat_subscription.py", line 102, in _ansiballz_main() File "/root/.ansible/tmp/ansible-tmp-1762629769.5202293-10278-20579478037084/AnsiballZ_redhat_subscription.py", line 94, in _ansiballz_main invoke_module(zipped_mod, temp_path, ANSIBALLZ_PARAMS) File "/root/.ansible/tmp/ansible-tmp-1762629769.5202293-10278-20579478037084/AnsiballZ_redhat_subscription.py", line 40, in invoke_module runpy.run_module(mod_name='ansible_collections.community.general.plugins.modules.redhat_subscription', init_globals=None, run_name='__main__', alter_sys=True) File "/usr/lib64/python3.6/runpy.py", line 201, in run_module mod_name, mod_spec, code = _get_module_details(mod_name) File "/usr/lib64/python3.6/runpy.py", line 128, in _get_module_details spec = importlib.util.find_spec(mod_name) File "/usr/lib64/python3.6/importlib/util.py", line 89, in find_spec return _find_spec(fullname, parent.__path__) File "", line 894, in _find_spec File "", line 1157, in find_spec File "", line 1131, in _get_spec File "", line 1112, in _legacy_get_spec File "", line 441, in spec_from_loader File "", line 544, in spec_from_file_location File "/tmp/ansible_community.general.redhat_subscription_payload_91opkzs5/ansible_community.general.redhat_subscription_payload.zip/ansible_collections/community/general/plugins/modules/redhat_subscription.py", line 8 SyntaxError: future feature annotations is not defined MODULE_STDERR: Shared connection to 10.31.40.154 closed. TASK [Clean up Candlepin container] ******************************************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tests_environments.yml:111 Saturday 08 November 2025 14:22:49 -0500 (0:00:00.348) 0:01:58.670 ***** included: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/teardown_candlepin.yml for managed-node2 TASK [Check if the candlepin container exists] ********************************* task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/teardown_candlepin.yml:6 Saturday 08 November 2025 14:22:49 -0500 (0:00:00.020) 0:01:58.690 ***** ok: [managed-node2] => { "changed": false, "cmd": [ "podman", "ps", "-a", "--filter", "name=candlepin" ], "delta": "0:00:00.045650", "end": "2025-11-08 14:22:50.154996", "rc": 0, "start": "2025-11-08 14:22:50.109346" } STDOUT: CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES b9ded374cfc8 ghcr.io/candlepin/candlepin-unofficial:latest /sbin/init 39 seconds ago Up 39 seconds 0.0.0.0:8080->8080/tcp, 0.0.0.0:8443->8443/tcp candlepin TASK [Ensure that Candlepin container doesn't exist] *************************** task path: /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/teardown_candlepin.yml:17 Saturday 08 November 2025 14:22:50 -0500 (0:00:00.383) 0:01:59.074 ***** changed: [managed-node2] => { "changed": true, "cmd": [ "podman", "stop", "candlepin" ], "delta": "0:00:00.805971", "end": "2025-11-08 14:22:51.301014", "rc": 0, "start": "2025-11-08 14:22:50.495043" } STDOUT: candlepin PLAY RECAP ********************************************************************* managed-node2 : ok=34 changed=6 unreachable=0 failed=2 skipped=22 rescued=1 ignored=0 SYSTEM ROLES ERRORS BEGIN v1 [ { "ansible_version": "2.9.27", "end_time": "2025-11-08T19:22:45.980879+00:00Z", "host": "managed-node2", "message": "MODULE FAILURE\nSee stdout/stderr for the exact error", "rc": 1, "start_time": "2025-11-08T19:22:45.509022+00:00Z", "task_name": "Call subscription-manager", "task_path": "/tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/subscription-manager.yml:23" }, { "ansible_version": "2.9.27", "end_time": "2025-11-08T19:22:49.313578+00:00Z", "host": "managed-node2", "message": "MODULE FAILURE\nSee stdout/stderr for the exact error", "rc": 1, "start_time": "2025-11-08T19:22:48.951901+00:00Z", "task_name": "Call subscription-manager", "task_path": "/tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/subscription-manager.yml:23" }, { "ansible_version": "2.9.27", "end_time": "2025-11-08T19:22:49.805487+00:00Z", "host": "managed-node2", "message": "MODULE FAILURE\nSee stdout/stderr for the exact error", "rc": 1, "start_time": "2025-11-08T19:22:49.459825+00:00Z", "task_name": "Call subscription-manager", "task_path": "/tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/subscription-manager.yml:23" } ] SYSTEM ROLES ERRORS END v1 TASKS RECAP ******************************************************************** Saturday 08 November 2025 14:22:51 -0500 (0:00:01.152) 0:02:00.227 ***** =============================================================================== Install needed packages ------------------------------------------------ 60.72s /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:31 Start Candlepin container ---------------------------------------------- 15.93s /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:41 Wait for started Candlepin --------------------------------------------- 14.57s /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:108 fedora.linux_system_roles.rhc : Ensure required packages are installed --- 4.51s /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/subscription-manager.yml:3 Ensure directories exist ------------------------------------------------ 3.77s /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:62 fedora.linux_system_roles.rhc : Ensure required packages are installed --- 2.82s /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/subscription-manager.yml:3 Copy product certificates ----------------------------------------------- 2.55s /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:72 Update system certificates store ---------------------------------------- 2.12s /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:101 Check if the candlepin container exists --------------------------------- 2.03s /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/teardown_candlepin.yml:6 Add environments -------------------------------------------------------- 1.66s /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:127 Copy Candlepin CA certificate for subscription-manager ------------------ 1.43s /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:83 Copy Candlepin CA certificate for system -------------------------------- 1.35s /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:92 Ensure that Candlepin container doesn't exist --------------------------- 1.15s /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/teardown_candlepin.yml:17 Check if system is ostree ----------------------------------------------- 0.66s /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_test_data.yml:32 Get facts for external test data ---------------------------------------- 0.62s /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:9 Install GPG key for RPM repositories ------------------------------------ 0.54s /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:118 Check Candlepin works --------------------------------------------------- 0.51s /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/check_candlepin.yml:3 fedora.linux_system_roles.rhc : Call subscription-manager --------------- 0.48s /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/subscription-manager.yml:23 Add candlepin hostname to /etc/hosts ------------------------------------ 0.44s /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:25 Ensure ansible_facts used by the test ----------------------------------- 0.42s /tmp/collections-O5T/ansible_collections/fedora/linux_system_roles/tests/rhc/tests_environments.yml:24 -- Logs begin at Sat 2025-11-08 14:16:40 EST, end at Sat 2025-11-08 14:22:51 EST. -- Nov 08 14:20:51 managed-node2 sshd[7083]: Accepted publickey for root from 10.31.42.172 port 51280 ssh2: ECDSA SHA256:7tV+4r/1rRCam0ILhv5aWnO3ErrXak3pFqIPmQYx2zc Nov 08 14:20:51 managed-node2 systemd-logind[593]: New session 8 of user root. -- Subject: A new session 8 has been created for user root -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A new session with the ID 8 has been created for the user root. -- -- The leading process of the session is 7083. Nov 08 14:20:51 managed-node2 systemd[1]: Started Session 8 of user root. -- Subject: Unit session-8.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit session-8.scope has finished starting up. -- -- The start-up result is done. Nov 08 14:20:51 managed-node2 sshd[7083]: pam_unix(sshd:session): session opened for user root by (uid=0) Nov 08 14:20:51 managed-node2 sudo[7228]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-bkrkmwzzcdhnryopixnoirqbdbcejedh ; /usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1762629651.4841216-7691-127302758885837/AnsiballZ_stat.py' Nov 08 14:20:51 managed-node2 sudo[7228]: pam_unix(sudo:session): session opened for user root by root(uid=0) Nov 08 14:20:51 managed-node2 platform-python[7231]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Nov 08 14:20:51 managed-node2 sudo[7228]: pam_unix(sudo:session): session closed for user root Nov 08 14:20:52 managed-node2 sudo[7354]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-trqtledajgwittvuplfyfjutvrooscrj ; /usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1762629652.053395-7708-98833313087571/AnsiballZ_setup.py' Nov 08 14:20:52 managed-node2 sudo[7354]: pam_unix(sudo:session): session opened for user root by root(uid=0) Nov 08 14:20:52 managed-node2 platform-python[7357]: ansible-setup Invoked with gather_subset=['!all', '!min', 'distribution', 'distribution_major_version'] gather_timeout=10 filter=* fact_path=/etc/ansible/facts.d Nov 08 14:20:52 managed-node2 sudo[7354]: pam_unix(sudo:session): session closed for user root Nov 08 14:20:52 managed-node2 sudo[7484]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-dsnuxsxemstnmdhooypbcdjhzfwudrvw ; /usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1762629652.7580009-7728-244178281915362/AnsiballZ_lineinfile.py' Nov 08 14:20:52 managed-node2 sudo[7484]: pam_unix(sudo:session): session opened for user root by root(uid=0) Nov 08 14:20:53 managed-node2 platform-python[7487]: ansible-lineinfile Invoked with path=/etc/hosts line=127.0.0.1 candlepin.local regexp=.*candlepin.local state=present backrefs=False create=False backup=False firstmatch=False follow=False unsafe_writes=False insertafter=None insertbefore=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None src=None force=None content=NOT_LOGGING_PARAMETER remote_src=None delimiter=None directory_mode=None Nov 08 14:20:53 managed-node2 sudo[7484]: pam_unix(sudo:session): session closed for user root Nov 08 14:20:53 managed-node2 sudo[7610]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-fklfyizpgmypbmlutllbrcupknjqdyxh ; /usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1762629653.1868305-7737-240882856740948/AnsiballZ_setup.py' Nov 08 14:20:53 managed-node2 sudo[7610]: pam_unix(sudo:session): session opened for user root by root(uid=0) Nov 08 14:20:53 managed-node2 platform-python[7613]: ansible-setup Invoked with filter=ansible_pkg_mgr gather_subset=['!all'] gather_timeout=10 fact_path=/etc/ansible/facts.d Nov 08 14:20:53 managed-node2 sudo[7610]: pam_unix(sudo:session): session closed for user root Nov 08 14:20:53 managed-node2 sudo[7681]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-jixsobcqjbejuqbmiyrohhexekyxsxgs ; /usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1762629653.1868305-7737-240882856740948/AnsiballZ_dnf.py' Nov 08 14:20:53 managed-node2 sudo[7681]: pam_unix(sudo:session): session opened for user root by root(uid=0) Nov 08 14:20:54 managed-node2 platform-python[7684]: ansible-dnf Invoked with name=['podman'] state=present allow_downgrade=False autoremove=False bugfix=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True lock_timeout=30 conf_file=None disable_excludes=None download_dir=None list=None releasever=None Nov 08 14:21:10 managed-node2 dbus-daemon[595]: [system] Reloaded configuration Nov 08 14:21:11 managed-node2 setsebool[7763]: The virt_use_nfs policy boolean was changed to 1 by root Nov 08 14:21:11 managed-node2 setsebool[7763]: The virt_sandbox_use_all_caps policy boolean was changed to 1 by root Nov 08 14:21:26 managed-node2 kernel: SELinux: Converting 360 SID table entries... Nov 08 14:21:26 managed-node2 kernel: SELinux: policy capability network_peer_controls=1 Nov 08 14:21:26 managed-node2 kernel: SELinux: policy capability open_perms=1 Nov 08 14:21:26 managed-node2 kernel: SELinux: policy capability extended_socket_class=1 Nov 08 14:21:26 managed-node2 kernel: SELinux: policy capability always_check_network=0 Nov 08 14:21:26 managed-node2 kernel: SELinux: policy capability cgroup_seclabel=1 Nov 08 14:21:26 managed-node2 kernel: SELinux: policy capability nnp_nosuid_transition=1 Nov 08 14:21:26 managed-node2 dbus-daemon[595]: [system] Reloaded configuration Nov 08 14:21:27 managed-node2 kernel: fuse: init (API version 7.34) Nov 08 14:21:27 managed-node2 systemd[1]: Mounting FUSE Control File System... -- Subject: Unit sys-fs-fuse-connections.mount has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit sys-fs-fuse-connections.mount has begun starting up. Nov 08 14:21:27 managed-node2 systemd[1]: Mounted FUSE Control File System. -- Subject: Unit sys-fs-fuse-connections.mount has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit sys-fs-fuse-connections.mount has finished starting up. -- -- The start-up result is done. Nov 08 14:21:28 managed-node2 dbus-daemon[595]: [system] Reloaded configuration Nov 08 14:21:28 managed-node2 dbus-daemon[595]: [system] Reloaded configuration Nov 08 14:21:52 managed-node2 systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. -- Subject: Unit run-rec2340f435d94a548bed15235102f58a.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit run-rec2340f435d94a548bed15235102f58a.service has finished starting up. -- -- The start-up result is done. Nov 08 14:21:52 managed-node2 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Nov 08 14:21:52 managed-node2 systemd[1]: Starting man-db-cache-update.service... -- Subject: Unit man-db-cache-update.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit man-db-cache-update.service has begun starting up. Nov 08 14:21:52 managed-node2 systemd[1]: Reloading. Nov 08 14:21:53 managed-node2 sudo[7681]: pam_unix(sudo:session): session closed for user root Nov 08 14:21:54 managed-node2 systemd[1]: man-db-cache-update.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit man-db-cache-update.service has successfully entered the 'dead' state. Nov 08 14:21:54 managed-node2 systemd[1]: Started man-db-cache-update.service. -- Subject: Unit man-db-cache-update.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit man-db-cache-update.service has finished starting up. -- -- The start-up result is done. Nov 08 14:21:54 managed-node2 systemd[1]: run-rec2340f435d94a548bed15235102f58a.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit run-rec2340f435d94a548bed15235102f58a.service has successfully entered the 'dead' state. Nov 08 14:21:54 managed-node2 sudo[10239]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-wjmvrxtenbryaqproaxvnpizhpqjpttx ; /usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1762629713.9495342-9579-240748523807549/AnsiballZ_command.py' Nov 08 14:21:54 managed-node2 sudo[10239]: pam_unix(sudo:session): session opened for user root by root(uid=0) Nov 08 14:21:54 managed-node2 platform-python[10242]: ansible-command Invoked with argv=['podman', 'ps', '-a', '--filter', 'name=candlepin'] warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True _raw_params=None chdir=None executable=None creates=None removes=None stdin=None Nov 08 14:21:55 managed-node2 kernel: evm: overlay not supported Nov 08 14:21:55 managed-node2 systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Nov 08 14:21:55 managed-node2 sudo[10239]: pam_unix(sudo:session): session closed for user root Nov 08 14:21:56 managed-node2 sudo[10374]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-qvenxjieevambvihkmcphrzcfximukvd ; /usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1762629716.0109446-9604-159558001171457/AnsiballZ_command.py' Nov 08 14:21:56 managed-node2 sudo[10374]: pam_unix(sudo:session): session opened for user root by root(uid=0) Nov 08 14:21:56 managed-node2 platform-python[10377]: ansible-command Invoked with argv=['podman', 'run', '--rm', '--detach', '--hostname', 'candlepin.local', '--name', 'candlepin', '--publish', '8443:8443', '--publish', '8080:8080', 'ghcr.io/candlepin/candlepin-unofficial'] warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True _raw_params=None chdir=None executable=None creates=None removes=None stdin=None Nov 08 14:22:10 managed-node2 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Nov 08 14:22:10 managed-node2 NetworkManager[674]: [1762629730.9734] manager: (cni-podman0): new Bridge device (/org/freedesktop/NetworkManager/Devices/3) Nov 08 14:22:10 managed-node2 systemd-udevd[10407]: Using default interface naming scheme 'rhel-8.0'. Nov 08 14:22:10 managed-node2 systemd-udevd[10407]: link_config: autonegotiation is unset or enabled, the speed and duplex are not writable. Nov 08 14:22:10 managed-node2 systemd-udevd[10407]: Could not generate persistent MAC address for cni-podman0: No such file or directory Nov 08 14:22:10 managed-node2 NetworkManager[674]: [1762629730.9950] manager: (veth4ebd24c7): new Veth device (/org/freedesktop/NetworkManager/Devices/4) Nov 08 14:22:10 managed-node2 kernel: IPv6: ADDRCONF(NETDEV_UP): veth4ebd24c7: link is not ready Nov 08 14:22:10 managed-node2 kernel: cni-podman0: port 1(veth4ebd24c7) entered blocking state Nov 08 14:22:10 managed-node2 kernel: cni-podman0: port 1(veth4ebd24c7) entered disabled state Nov 08 14:22:10 managed-node2 kernel: device veth4ebd24c7 entered promiscuous mode Nov 08 14:22:10 managed-node2 NetworkManager[674]: [1762629730.9972] device (cni-podman0): state change: unmanaged -> unavailable (reason 'connection-assumed', sys-iface-state: 'external') Nov 08 14:22:10 managed-node2 NetworkManager[674]: [1762629730.9975] device (cni-podman0): state change: unavailable -> disconnected (reason 'connection-assumed', sys-iface-state: 'external') Nov 08 14:22:10 managed-node2 NetworkManager[674]: [1762629730.9981] device (cni-podman0): Activation: starting connection 'cni-podman0' (b9ac5d76-640b-4988-afe1-bddd93f3b573) Nov 08 14:22:10 managed-node2 NetworkManager[674]: [1762629730.9982] device (cni-podman0): state change: disconnected -> prepare (reason 'none', sys-iface-state: 'external') Nov 08 14:22:10 managed-node2 NetworkManager[674]: [1762629730.9985] device (cni-podman0): state change: prepare -> config (reason 'none', sys-iface-state: 'external') Nov 08 14:22:10 managed-node2 NetworkManager[674]: [1762629730.9986] device (cni-podman0): state change: config -> ip-config (reason 'none', sys-iface-state: 'external') Nov 08 14:22:10 managed-node2 NetworkManager[674]: [1762629730.9988] device (cni-podman0): state change: ip-config -> ip-check (reason 'none', sys-iface-state: 'external') Nov 08 14:22:11 managed-node2 dbus-daemon[595]: [system] Activating via systemd: service name='org.freedesktop.nm_dispatcher' unit='dbus-org.freedesktop.nm-dispatcher.service' requested by ':1.5' (uid=0 pid=674 comm="/usr/sbin/NetworkManager --no-daemon " label="system_u:system_r:NetworkManager_t:s0") Nov 08 14:22:11 managed-node2 systemd-udevd[10407]: link_config: autonegotiation is unset or enabled, the speed and duplex are not writable. Nov 08 14:22:11 managed-node2 systemd-udevd[10407]: Could not generate persistent MAC address for veth4ebd24c7: No such file or directory Nov 08 14:22:11 managed-node2 systemd[1]: Starting Network Manager Script Dispatcher Service... -- Subject: Unit NetworkManager-dispatcher.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit NetworkManager-dispatcher.service has begun starting up. Nov 08 14:22:11 managed-node2 kernel: IPv6: ADDRCONF(NETDEV_UP): eth0: link is not ready Nov 08 14:22:11 managed-node2 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready Nov 08 14:22:11 managed-node2 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): veth4ebd24c7: link becomes ready Nov 08 14:22:11 managed-node2 kernel: cni-podman0: port 1(veth4ebd24c7) entered blocking state Nov 08 14:22:11 managed-node2 kernel: cni-podman0: port 1(veth4ebd24c7) entered forwarding state Nov 08 14:22:11 managed-node2 NetworkManager[674]: [1762629731.0317] device (veth4ebd24c7): carrier: link connected Nov 08 14:22:11 managed-node2 NetworkManager[674]: [1762629731.0320] device (cni-podman0): carrier: link connected Nov 08 14:22:11 managed-node2 dbus-daemon[595]: [system] Successfully activated service 'org.freedesktop.nm_dispatcher' Nov 08 14:22:11 managed-node2 systemd[1]: Started Network Manager Script Dispatcher Service. -- Subject: Unit NetworkManager-dispatcher.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit NetworkManager-dispatcher.service has finished starting up. -- -- The start-up result is done. Nov 08 14:22:11 managed-node2 NetworkManager[674]: [1762629731.0593] device (cni-podman0): state change: ip-check -> secondaries (reason 'none', sys-iface-state: 'external') Nov 08 14:22:11 managed-node2 NetworkManager[674]: [1762629731.0595] device (cni-podman0): state change: secondaries -> activated (reason 'none', sys-iface-state: 'external') Nov 08 14:22:11 managed-node2 NetworkManager[674]: [1762629731.0599] device (cni-podman0): Activation: successful, device activated. Nov 08 14:22:11 managed-node2 systemd[1]: Created slice machine.slice. -- Subject: Unit machine.slice has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit machine.slice has finished starting up. -- -- The start-up result is done. Nov 08 14:22:11 managed-node2 systemd[1]: Started libpod-conmon-b9ded374cfc89cf2a47688a5a1eee040580bdb4398a1ca54f42d1325a4e44440.scope. -- Subject: Unit libpod-conmon-b9ded374cfc89cf2a47688a5a1eee040580bdb4398a1ca54f42d1325a4e44440.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-conmon-b9ded374cfc89cf2a47688a5a1eee040580bdb4398a1ca54f42d1325a4e44440.scope has finished starting up. -- -- The start-up result is done. Nov 08 14:22:11 managed-node2 systemd[1]: Started libcontainer container b9ded374cfc89cf2a47688a5a1eee040580bdb4398a1ca54f42d1325a4e44440. -- Subject: Unit libpod-b9ded374cfc89cf2a47688a5a1eee040580bdb4398a1ca54f42d1325a4e44440.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit libpod-b9ded374cfc89cf2a47688a5a1eee040580bdb4398a1ca54f42d1325a4e44440.scope has finished starting up. -- -- The start-up result is done. Nov 08 14:22:11 managed-node2 systemd[1]: run-runc-b9ded374cfc89cf2a47688a5a1eee040580bdb4398a1ca54f42d1325a4e44440-runc.ZSq3j3.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit run-runc-b9ded374cfc89cf2a47688a5a1eee040580bdb4398a1ca54f42d1325a4e44440-runc.ZSq3j3.mount has successfully entered the 'dead' state. Nov 08 14:22:11 managed-node2 sudo[10374]: pam_unix(sudo:session): session closed for user root Nov 08 14:22:12 managed-node2 sudo[10783]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-hkomnkqzkoukfzswoymnkwqzwumtazca ; /usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1762629732.057532-9725-127652510896640/AnsiballZ_file.py' Nov 08 14:22:12 managed-node2 sudo[10783]: pam_unix(sudo:session): session opened for user root by root(uid=0) Nov 08 14:22:12 managed-node2 platform-python[10786]: ansible-file Invoked with path=/etc/pki/product state=directory mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Nov 08 14:22:12 managed-node2 sudo[10783]: pam_unix(sudo:session): session closed for user root Nov 08 14:22:13 managed-node2 sudo[10909]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-httaiyakpuhqsqhmndwacdlydpeenujc ; /usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1762629733.1226351-9725-182460111923402/AnsiballZ_file.py' Nov 08 14:22:13 managed-node2 sudo[10909]: pam_unix(sudo:session): session opened for user root by root(uid=0) Nov 08 14:22:13 managed-node2 platform-python[10912]: ansible-file Invoked with path=/etc/pki/product-default state=directory mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Nov 08 14:22:13 managed-node2 sudo[10909]: pam_unix(sudo:session): session closed for user root Nov 08 14:22:15 managed-node2 sudo[11035]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-lqskwhzluyblblwlngjvsrckquptovqf ; /usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1762629734.0575988-9725-44985681757312/AnsiballZ_file.py' Nov 08 14:22:15 managed-node2 sudo[11035]: pam_unix(sudo:session): session opened for user root by root(uid=0) Nov 08 14:22:15 managed-node2 platform-python[11038]: ansible-file Invoked with path=/etc/rhsm/ca state=directory mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Nov 08 14:22:15 managed-node2 sudo[11035]: pam_unix(sudo:session): session closed for user root Nov 08 14:22:17 managed-node2 sudo[11161]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-nzsfhxogjedhpyrzeeuklekwfvazlhsd ; /usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1762629735.8702996-9769-273853506748715/AnsiballZ_command.py' Nov 08 14:22:17 managed-node2 sudo[11161]: pam_unix(sudo:session): session opened for user root by root(uid=0) Nov 08 14:22:17 managed-node2 platform-python[11164]: ansible-command Invoked with argv=['podman', 'cp', 'candlepin:/home/candlepin/devel/candlepin/generated_certs/7050.pem', '/etc/pki/product-default/'] warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True _raw_params=None chdir=None executable=None creates=None removes=None stdin=None Nov 08 14:22:18 managed-node2 sudo[11161]: pam_unix(sudo:session): session closed for user root Nov 08 14:22:18 managed-node2 sudo[11322]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-kjhpnkajoirbexpyxbmatlmoxnrphmbi ; /usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1762629738.4633725-9803-105104645588464/AnsiballZ_command.py' Nov 08 14:22:18 managed-node2 sudo[11322]: pam_unix(sudo:session): session opened for user root by root(uid=0) Nov 08 14:22:19 managed-node2 platform-python[11325]: ansible-command Invoked with argv=['podman', 'cp', 'candlepin:/etc/candlepin/certs/candlepin-ca.crt', '/etc/rhsm/ca/candlepin-ca.pem'] warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True _raw_params=None chdir=None executable=None creates=None removes=None stdin=None Nov 08 14:22:19 managed-node2 sudo[11322]: pam_unix(sudo:session): session closed for user root Nov 08 14:22:20 managed-node2 sudo[11485]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-pvjorqdmvivxznwyciekieatfurtzhit ; /usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1762629739.758624-9826-156023135163312/AnsiballZ_command.py' Nov 08 14:22:20 managed-node2 sudo[11485]: pam_unix(sudo:session): session opened for user root by root(uid=0) Nov 08 14:22:20 managed-node2 platform-python[11488]: ansible-command Invoked with argv=['podman', 'cp', 'candlepin:/etc/candlepin/certs/candlepin-ca.crt', '/etc/pki/ca-trust/source/anchors/candlepin-ca.pem'] warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True _raw_params=None chdir=None executable=None creates=None removes=None stdin=None Nov 08 14:22:20 managed-node2 sudo[11485]: pam_unix(sudo:session): session closed for user root Nov 08 14:22:21 managed-node2 systemd[1]: NetworkManager-dispatcher.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit NetworkManager-dispatcher.service has successfully entered the 'dead' state. Nov 08 14:22:21 managed-node2 sudo[11651]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-imkyeaayjjbzvodmsvwzhtkqdmfdnphi ; /usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1762629741.1934419-9860-268438231211876/AnsiballZ_command.py' Nov 08 14:22:21 managed-node2 sudo[11651]: pam_unix(sudo:session): session opened for user root by root(uid=0) Nov 08 14:22:21 managed-node2 platform-python[11654]: ansible-command Invoked with argv=['update-ca-trust', 'extract'] warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True _raw_params=None chdir=None executable=None creates=None removes=None stdin=None Nov 08 14:22:23 managed-node2 sudo[11651]: pam_unix(sudo:session): session closed for user root Nov 08 14:22:24 managed-node2 sudo[11786]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-dcujqdnlhjklwdkpbpqxszacdgcvobae ; /usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1762629743.3377252-9883-277733014059210/AnsiballZ_uri.py' Nov 08 14:22:24 managed-node2 sudo[11786]: pam_unix(sudo:session): session opened for user root by root(uid=0) Nov 08 14:22:24 managed-node2 platform-python[11789]: ansible-uri Invoked with url=https://candlepin.local:8443/candlepin method=HEAD validate_certs=False force=False http_agent=ansible-httpget use_proxy=True force_basic_auth=False body_format=raw return_content=False follow_redirects=safe status_code=[200] timeout=30 headers={} follow=False unsafe_writes=False url_username=None url_password=NOT_LOGGING_PARAMETER client_cert=None client_key=None dest=None body=None src=None creates=None removes=None unix_socket=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Nov 08 14:22:37 managed-node2 sudo[11786]: pam_unix(sudo:session): session closed for user root Nov 08 14:22:37 managed-node2 sudo[12008]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-qexpwufllveyhfeobkimgomsvonjnlaf ; /usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1762629757.7391105-9999-41635758179647/AnsiballZ_get_url.py' Nov 08 14:22:37 managed-node2 sudo[12008]: pam_unix(sudo:session): session opened for user root by root(uid=0) Nov 08 14:22:38 managed-node2 platform-python[12011]: ansible-get_url Invoked with url=http://candlepin.local:8080/RPM-GPG-KEY-candlepin dest=/etc/pki/rpm-gpg/RPM-GPG-KEY-candlepin mode=0644 force=False http_agent=ansible-httpget use_proxy=True validate_certs=True force_basic_auth=False sha256sum= checksum= timeout=10 follow=False unsafe_writes=False url_username=None url_password=NOT_LOGGING_PARAMETER client_cert=None client_key=None backup=None headers=None tmp_dest=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None src=None content=NOT_LOGGING_PARAMETER remote_src=None regexp=None delimiter=None directory_mode=None Nov 08 14:22:38 managed-node2 sudo[12008]: pam_unix(sudo:session): session closed for user root Nov 08 14:22:38 managed-node2 sudo[12134]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-oubswsthwyyujgztgkielhiikqdapfjt ; /usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1762629758.2820435-10021-34444680874780/AnsiballZ_uri.py' Nov 08 14:22:38 managed-node2 sudo[12134]: pam_unix(sudo:session): session opened for user root by root(uid=0) Nov 08 14:22:38 managed-node2 platform-python[12137]: ansible-uri Invoked with url=https://candlepin.local:8443/candlepin/owners/donaldduck/environments method=POST url_username=doc url_********=NOT_LOGGING_PARAMETER body_format=json body={'name': 'Environment 1', 'description': 'The environment 1', 'id': 'envId1'} force=False http_agent=ansible-httpget use_proxy=True validate_certs=True force_basic_auth=False return_content=False follow_redirects=safe status_code=[200] timeout=30 headers={} follow=False unsafe_writes=False client_cert=None client_key=None dest=None src=None creates=None removes=None unix_socket=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Nov 08 14:22:39 managed-node2 sudo[12134]: pam_unix(sudo:session): session closed for user root Nov 08 14:22:39 managed-node2 sudo[12262]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-vxzxiofjbzyemijcyrjmxufncjnhkjkm ; /usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1762629759.314413-10021-187781414442078/AnsiballZ_uri.py' Nov 08 14:22:39 managed-node2 sudo[12262]: pam_unix(sudo:session): session opened for user root by root(uid=0) Nov 08 14:22:39 managed-node2 platform-python[12265]: ansible-uri Invoked with url=https://candlepin.local:8443/candlepin/owners/donaldduck/environments method=POST url_username=doc url_********=NOT_LOGGING_PARAMETER body_format=json body={'name': 'Environment 2', 'description': 'The environment 2', 'id': 'envId2'} force=False http_agent=ansible-httpget use_proxy=True validate_certs=True force_basic_auth=False return_content=False follow_redirects=safe status_code=[200] timeout=30 headers={} follow=False unsafe_writes=False client_cert=None client_key=None dest=None src=None creates=None removes=None unix_socket=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Nov 08 14:22:39 managed-node2 sudo[12262]: pam_unix(sudo:session): session closed for user root Nov 08 14:22:40 managed-node2 sudo[12390]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-wvzpqkgtviazziuysghhrygjhnwtfqsq ; /usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1762629759.9270391-10052-177225179440700/AnsiballZ_uri.py' Nov 08 14:22:40 managed-node2 sudo[12390]: pam_unix(sudo:session): session opened for user root by root(uid=0) Nov 08 14:22:40 managed-node2 platform-python[12393]: ansible-uri Invoked with url=https://candlepin.local:8443/candlepin method=HEAD validate_certs=False force=False http_agent=ansible-httpget use_proxy=True force_basic_auth=False body_format=raw return_content=False follow_redirects=safe status_code=[200] timeout=30 headers={} follow=False unsafe_writes=False url_username=None url_password=NOT_LOGGING_PARAMETER client_cert=None client_key=None dest=None body=None src=None creates=None removes=None unix_socket=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None content=NOT_LOGGING_PARAMETER backup=None remote_src=None regexp=None delimiter=None directory_mode=None Nov 08 14:22:40 managed-node2 sudo[12390]: pam_unix(sudo:session): session closed for user root Nov 08 14:22:40 managed-node2 sudo[12516]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-pbeuqlpmpxjroaztrfsckwxjpfboxrhc ; /usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1762629760.4129753-10069-25875713861391/AnsiballZ_setup.py' Nov 08 14:22:40 managed-node2 sudo[12516]: pam_unix(sudo:session): session opened for user root by root(uid=0) Nov 08 14:22:40 managed-node2 platform-python[12519]: ansible-setup Invoked with gather_subset=['!all', '!min', 'distribution', 'distribution_major_version', 'distribution_version'] gather_timeout=10 filter=* fact_path=/etc/ansible/facts.d Nov 08 14:22:40 managed-node2 sudo[12516]: pam_unix(sudo:session): session closed for user root Nov 08 14:22:41 managed-node2 sudo[12646]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-hkzbbpaiogblwsssjkfghcatmjxofunx ; /usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1762629761.021437-10097-122711510719924/AnsiballZ_setup.py' Nov 08 14:22:41 managed-node2 sudo[12646]: pam_unix(sudo:session): session opened for user root by root(uid=0) Nov 08 14:22:41 managed-node2 platform-python[12649]: ansible-setup Invoked with filter=ansible_pkg_mgr gather_subset=['!all'] gather_timeout=10 fact_path=/etc/ansible/facts.d Nov 08 14:22:41 managed-node2 sudo[12646]: pam_unix(sudo:session): session closed for user root Nov 08 14:22:41 managed-node2 sudo[12717]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-ycxkstbleqjxnuvixmcmmdlatzweofzn ; /usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1762629761.021437-10097-122711510719924/AnsiballZ_dnf.py' Nov 08 14:22:41 managed-node2 sudo[12717]: pam_unix(sudo:session): session opened for user root by root(uid=0) Nov 08 14:22:41 managed-node2 platform-python[12720]: ansible-dnf Invoked with name=['subscription-manager'] state=present allow_downgrade=False autoremove=False bugfix=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True lock_timeout=30 conf_file=None disable_excludes=None download_dir=None list=None releasever=None Nov 08 14:22:44 managed-node2 dbus-daemon[595]: [system] Reloaded configuration Nov 08 14:22:44 managed-node2 dbus-daemon[595]: [system] Reloaded configuration Nov 08 14:22:44 managed-node2 dbus-daemon[595]: [system] Reloaded configuration Nov 08 14:22:44 managed-node2 dbus-daemon[595]: [system] Reloaded configuration Nov 08 14:22:44 managed-node2 dbus-daemon[595]: [system] Reloaded configuration Nov 08 14:22:44 managed-node2 dbus-daemon[595]: [system] Reloaded configuration Nov 08 14:22:44 managed-node2 dbus-daemon[595]: [system] Reloaded configuration Nov 08 14:22:44 managed-node2 systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. -- Subject: Unit run-r51edd561a16d41f98ba15f10f63b6cd6.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit run-r51edd561a16d41f98ba15f10f63b6cd6.service has finished starting up. -- -- The start-up result is done. Nov 08 14:22:44 managed-node2 systemd[1]: Starting man-db-cache-update.service... -- Subject: Unit man-db-cache-update.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit man-db-cache-update.service has begun starting up. Nov 08 14:22:44 managed-node2 systemd[1]: Reloading. Nov 08 14:22:45 managed-node2 systemd[1]: man-db-cache-update.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit man-db-cache-update.service has successfully entered the 'dead' state. Nov 08 14:22:45 managed-node2 systemd[1]: Started man-db-cache-update.service. -- Subject: Unit man-db-cache-update.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit man-db-cache-update.service has finished starting up. -- -- The start-up result is done. Nov 08 14:22:45 managed-node2 systemd[1]: run-r51edd561a16d41f98ba15f10f63b6cd6.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit run-r51edd561a16d41f98ba15f10f63b6cd6.service has successfully entered the 'dead' state. Nov 08 14:22:45 managed-node2 sudo[12717]: pam_unix(sudo:session): session closed for user root Nov 08 14:22:45 managed-node2 sudo[12985]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-cgcdusssnwrvuhpsldgaswyjhzjzpkpo ; /usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1762629765.5851321-10161-74016621386902/AnsiballZ_redhat_subscription.py' Nov 08 14:22:45 managed-node2 sudo[12985]: pam_unix(sudo:session): session opened for user root by root(uid=0) Nov 08 14:22:45 managed-node2 sudo[12985]: pam_unix(sudo:session): session closed for user root Nov 08 14:22:46 managed-node2 sudo[13111]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-rlidiiouaaveikxbapxugatdnnuvohnm ; /usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1762629766.159865-10192-119992344148103/AnsiballZ_setup.py' Nov 08 14:22:46 managed-node2 sudo[13111]: pam_unix(sudo:session): session opened for user root by root(uid=0) Nov 08 14:22:46 managed-node2 platform-python[13114]: ansible-setup Invoked with filter=ansible_pkg_mgr gather_subset=['!all'] gather_timeout=10 fact_path=/etc/ansible/facts.d Nov 08 14:22:46 managed-node2 sudo[13111]: pam_unix(sudo:session): session closed for user root Nov 08 14:22:46 managed-node2 sudo[13182]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-egdhqsxendrllnpykiejaavxbdjcxnca ; /usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1762629766.159865-10192-119992344148103/AnsiballZ_dnf.py' Nov 08 14:22:46 managed-node2 sudo[13182]: pam_unix(sudo:session): session opened for user root by root(uid=0) Nov 08 14:22:46 managed-node2 platform-python[13185]: ansible-dnf Invoked with name=['subscription-manager'] state=present allow_downgrade=False autoremove=False bugfix=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True lock_timeout=30 conf_file=None disable_excludes=None download_dir=None list=None releasever=None Nov 08 14:22:48 managed-node2 sudo[13182]: pam_unix(sudo:session): session closed for user root Nov 08 14:22:49 managed-node2 sudo[13308]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-ogqtetcrxlstidjygrvqbwmmbgwfpdcq ; /usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1762629769.027101-10242-824133678158/AnsiballZ_redhat_subscription.py' Nov 08 14:22:49 managed-node2 sudo[13308]: pam_unix(sudo:session): session opened for user root by root(uid=0) Nov 08 14:22:49 managed-node2 sudo[13308]: pam_unix(sudo:session): session closed for user root Nov 08 14:22:49 managed-node2 sudo[13434]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-rlvsavogivewsibckqlactmopmfjpzoc ; /usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1762629769.5202293-10278-20579478037084/AnsiballZ_redhat_subscription.py' Nov 08 14:22:49 managed-node2 sudo[13434]: pam_unix(sudo:session): session opened for user root by root(uid=0) Nov 08 14:22:49 managed-node2 sudo[13434]: pam_unix(sudo:session): session closed for user root Nov 08 14:22:49 managed-node2 sudo[13560]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-ixwkhfhodmddaekezfnumwdgwgwmukjf ; /usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1762629769.864288-10289-276692465636039/AnsiballZ_command.py' Nov 08 14:22:49 managed-node2 sudo[13560]: pam_unix(sudo:session): session opened for user root by root(uid=0) Nov 08 14:22:50 managed-node2 platform-python[13563]: ansible-command Invoked with argv=['podman', 'ps', '-a', '--filter', 'name=candlepin'] warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True _raw_params=None chdir=None executable=None creates=None removes=None stdin=None Nov 08 14:22:50 managed-node2 sudo[13560]: pam_unix(sudo:session): session closed for user root Nov 08 14:22:50 managed-node2 sudo[13693]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-oxkzgpuskmxhexsgwaybzsutkuxfsuqu ; /usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1762629770.2501838-10305-127304106217436/AnsiballZ_command.py' Nov 08 14:22:50 managed-node2 sudo[13693]: pam_unix(sudo:session): session opened for user root by root(uid=0) Nov 08 14:22:50 managed-node2 platform-python[13696]: ansible-command Invoked with argv=['podman', 'stop', 'candlepin'] warn=True _uses_shell=False stdin_add_newline=True strip_empty_ends=True _raw_params=None chdir=None executable=None creates=None removes=None stdin=None Nov 08 14:22:51 managed-node2 systemd[1]: libpod-b9ded374cfc89cf2a47688a5a1eee040580bdb4398a1ca54f42d1325a4e44440.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-b9ded374cfc89cf2a47688a5a1eee040580bdb4398a1ca54f42d1325a4e44440.scope has successfully entered the 'dead' state. Nov 08 14:22:51 managed-node2 systemd[1]: libpod-b9ded374cfc89cf2a47688a5a1eee040580bdb4398a1ca54f42d1325a4e44440.scope: Consumed 46.421s CPU time -- Subject: Resources consumed by unit runtime -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-b9ded374cfc89cf2a47688a5a1eee040580bdb4398a1ca54f42d1325a4e44440.scope completed and consumed the indicated resources. Nov 08 14:22:51 managed-node2 kernel: cni-podman0: port 1(veth4ebd24c7) entered disabled state Nov 08 14:22:51 managed-node2 kernel: device veth4ebd24c7 left promiscuous mode Nov 08 14:22:51 managed-node2 kernel: cni-podman0: port 1(veth4ebd24c7) entered disabled state Nov 08 14:22:51 managed-node2 systemd[1]: run-netns-netns\x2d6b433352\x2d986f\x2da62f\x2d2931\x2d102a13782bdb.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit run-netns-netns\x2d6b433352\x2d986f\x2da62f\x2d2931\x2d102a13782bdb.mount has successfully entered the 'dead' state. Nov 08 14:22:51 managed-node2 systemd[1]: var-lib-containers-storage-overlay\x2dcontainers-b9ded374cfc89cf2a47688a5a1eee040580bdb4398a1ca54f42d1325a4e44440-userdata-shm.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay\x2dcontainers-b9ded374cfc89cf2a47688a5a1eee040580bdb4398a1ca54f42d1325a4e44440-userdata-shm.mount has successfully entered the 'dead' state. Nov 08 14:22:51 managed-node2 systemd[1]: var-lib-containers-storage-overlay-ca5c3ac7dce5860a860f8dffc55f56973d1d23b94a138212d18c175c4fc4b884-merged.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay-ca5c3ac7dce5860a860f8dffc55f56973d1d23b94a138212d18c175c4fc4b884-merged.mount has successfully entered the 'dead' state. Nov 08 14:22:51 managed-node2 systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit var-lib-containers-storage-overlay.mount has successfully entered the 'dead' state. Nov 08 14:22:51 managed-node2 sudo[13693]: pam_unix(sudo:session): session closed for user root Nov 08 14:22:51 managed-node2 systemd[1]: libpod-conmon-b9ded374cfc89cf2a47688a5a1eee040580bdb4398a1ca54f42d1325a4e44440.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit libpod-conmon-b9ded374cfc89cf2a47688a5a1eee040580bdb4398a1ca54f42d1325a4e44440.scope has successfully entered the 'dead' state. Nov 08 14:22:51 managed-node2 sshd[13845]: Accepted publickey for root from 10.31.42.172 port 44834 ssh2: RSA SHA256:9j1blwt3wcrRiGYZQ7ZGu9axm3cDklH6/z4c+Ee8CzE Nov 08 14:22:51 managed-node2 systemd[1]: Started Session 9 of user root. -- Subject: Unit session-9.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit session-9.scope has finished starting up. -- -- The start-up result is done. Nov 08 14:22:51 managed-node2 systemd-logind[593]: New session 9 of user root. -- Subject: A new session 9 has been created for user root -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A new session with the ID 9 has been created for the user root. -- -- The leading process of the session is 13845. Nov 08 14:22:51 managed-node2 sshd[13845]: pam_unix(sshd:session): session opened for user root by (uid=0) Nov 08 14:22:51 managed-node2 sshd[13848]: Received disconnect from 10.31.42.172 port 44834:11: disconnected by user Nov 08 14:22:51 managed-node2 sshd[13848]: Disconnected from user root 10.31.42.172 port 44834 Nov 08 14:22:51 managed-node2 sshd[13845]: pam_unix(sshd:session): session closed for user root Nov 08 14:22:51 managed-node2 systemd[1]: session-9.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit session-9.scope has successfully entered the 'dead' state. Nov 08 14:22:51 managed-node2 systemd-logind[593]: Session 9 logged out. Waiting for processes to exit. Nov 08 14:22:51 managed-node2 systemd-logind[593]: Removed session 9. -- Subject: Session 9 has been terminated -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A session with the ID 9 has been terminated. Nov 08 14:22:51 managed-node2 sshd[13869]: Accepted publickey for root from 10.31.42.172 port 44836 ssh2: RSA SHA256:9j1blwt3wcrRiGYZQ7ZGu9axm3cDklH6/z4c+Ee8CzE Nov 08 14:22:51 managed-node2 systemd[1]: Started Session 10 of user root. -- Subject: Unit session-10.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit session-10.scope has finished starting up. -- -- The start-up result is done. Nov 08 14:22:51 managed-node2 systemd-logind[593]: New session 10 of user root. -- Subject: A new session 10 has been created for user root -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A new session with the ID 10 has been created for the user root. -- -- The leading process of the session is 13869. Nov 08 14:22:51 managed-node2 sshd[13869]: pam_unix(sshd:session): session opened for user root by (uid=0)